var/home/core/zuul-output/0000755000175000017500000000000015156214607014534 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015156226201015471 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000306547415156226117020277 0ustar corecoreO,ikubelet.lognc9r~DYd` \-Hږ%C{sg5݁ϑ)Ӄis$WU)X6]Eڤ펯_ˎ6_o#oVݏKf핷ox[o8W5-b6"οƼ>UWm׫Y_?|uݗ[y[L-V_pY_P-bXwûxwAۋt[~ _P^~&RY,yDy~z]/:oXx$%X"LADA@@tkޕf{5Wbx=@^J})K3x~JkwI|YowS˷jכ̶]/8 N Rm(of`\r\L>{Jm 0{vRFEb%]˜(O)X}d\UlxDJfw6xi1U2 c#FD?2SgafO3|,ejoLR3[ D HJP1Ub2i]$HU^L_cZ_:F9TJJ{,mvgL;: ԓ$aёdt6o[ .`:J ]HmS>v5gCh31 )Kh3i J1hG{aD4iӌçN/e] o;ijiBx_2dd$YLYG(#?%U? ` 17ׅwڋًM)$Fiqw7Gt7L"u 0V9c  ˹dvYļU[ Z.׿/h QZ*U1|t5wKOM6_Oފ?O1nԝG?ƥF%QV5pDVHwԡ/.2h{qۀK8yUOdssdMvw`21ɻ]/ƛ"@8(PN_,_0;_x+Vy<h\dN9:bġ7 -Pwȹl;M@v4If{5C/(\ Q] R['>v*;o57sp$3nC|]|[>ӸUKޥg9b2oII"9 1"6Dkſ~I=嚲W9ȝQEkT/*BR =v*.h4(^&-Wg̫b]OBEFδW~N 97;Zp0s]UIĀg)4 B^S4t; *퇄u p}du ~z/9آs;DPsif39HoN λC?; H^-¸oZ( +"@@%'0MtWG uIo1]ߔr TGGJ\ C.iTEZ{(¥:-³xlՐ0A_Fݗw)(c>/1:N3cl.:f 3 JJ5Z<{kJ_O{*Z8Y CEO+'HqZY PTUJ2dic3w ?YQgpa` Z_0΁?kMPc_Ԝ*΄Bs`kmJ?t 53@հ1hr}=5t;at 9:Ѥ߯R> kH&Y``zG,z҄R K&Nh c{A`O'd1*-B[aL"T 1dȂ0TJ#r)٧4!)'qOFz|&4@2ƭ1-RN%?i¸ `eH&MJ!&ᙢ(<<-ja0Tazkm{ GYә7}U>>a>Ҟҝ>Ϗ ,ȓw`E_d$Ə{(he NSfX1982THwnUC9fDx5X@O5OޔL<'Adp]{1DL^5"Ϧޙ`F}W5XDV7V5EE9esYYfiMOV i/ f>3VQ 7,oT tMK%\t=[ٹ:11:2`c J1bV_gɊ:+^͒V,~0{gj"A, rXr*0ngY.] <ʜ6 ;,9VPAHuŠկiw=m{> *nacԇ&~hb[nӉ>'݌6od y$P'BOTLl-9Ja [$3BV2DC4l!TO C*Mrii1f5 JA *#jv߿Imy%u LOL8fq CXReQP2$TbgK !)CGm`XYB[9% E*:`cBCIqC(1&b f]fNhdQvݸCVA/e.# Okx܍>М>ӗom$rۇnu~Y݇̇TIwӜ'=A7,Q)-,=1A sK|ۜLɽy]ʸEO<-YEqKzϢ \{>dDLF amK/0-Vb=SSO|k1A[|gbͧɇد;:X:@;afU=Sru CK >Y%LwM*9ƀZ@=q' $dߋ{Ny[$ {ɴ6hOI']dC5`t9:GO: FmlN*:g^;T^B0$B=aK`3CmF1K>*Mk{_'֜dw$FEc* A>{avdt)8|mg定TN7,TEV |ɧ<Ғ8_iqEGPVC P2EU:F4!ʢlQHZ9E CBU)Y(S8)c yO[E}Lc&l .U轋RQ'Vt3,F3,#Y3,kJ3,LhVnKauomˠ_~g,ZByXϯ&Ksg3["66hŢFD&iQCFd4%h}̗Uɾ?si&2"C]uG~^`X\u6|6rcIF3b9O:j 2IN…D% Y"O j\WFI#R޸B4rc\>1yFX09'A%bDb0CPvw/T/ia v[|mAC5t[OM91c:VJR9&ksvJ;0ɝ$krogB= FYtЩOte=?>T&O{Ll)HClba1PIFĀ":tu^}.&R*!^pHPQuSVO$.KMb.:DK>WtWǭKv4@Va3"a`R@gbu%_J5Ґ 3DrH_HI\:U}UE$J @ٚeZE0(8ŋ ϓ{Bb$BrW XWz<%fpG"m%6PGEH^*JL֗J)oEv[Ң߃x[䚒}0BOnYr猸p$nu?ݣ RF]NHw2k혿q}lrCy u)xF$Z83Ec罋}[εUX%}< ݻln"sv&{b%^AAoۺ(I#hKD:Bߩ#蘈f=9oN*.Ѓ M#JC1?tean`3-SHq$2[ĜSjXRx?}-m6Mw'yR3q㕐)HW'X1BEb $xd(21i)//_і/Cޮm0VKz>I; >d[5Z=4>5!!T@[4 1.x XF`,?Hh]b-#3J( &uz u8.00-(9ŽZcX Jٯ^蒋*k.\MA/Xp9VqNo}#ƓOފgv[r*hy| IϭR-$$m!-W'wTi:4F5^z3/[{1LK[2nM|[<\t=3^qOp4y}|B}yu}뚬"P.ԘBn방u<#< A Q(j%e1!gkqiP(-ʢ-b7$66|*f\#ߍp{8sx[o%}wS`ýͽ>^U_S1VF20:d T2$47mSl*#lzFP_3yb.63>NKnJۦ^4*rB쑓:5Ǧ٨C.1`mU]+y_:,eXX맻c5ޖSwe݊O4L)69 War)|VϟT;Cq%KK-*i ѩQٰ`DݎGu( 꿢\cXn }7Ҫa nG{Y bcWa?\34 P U!7 _* kTuwmUr%ԀjƮĀdU#^ۈӕ3ΊeBO`^}ܖj49lnAvoI "%\;OF& wctغBܮl##mϸ.6p5k0C5PdKB g:=G<$w 24 6e/!~߽f)Q UbshY5mseڠ5_m4(sgz1v&YN2姟d4"?oWNW݃yh~%DTt^W7q.@ L⃳662G,:* $: e~7[/P%F on~$dƹɥO"dޢt|BpYqc@P`ڄj҆anCѢMU sf`Yɇك]@Rɯ?ٽf? ntպ$ˣ>TDNIGW .Z#YmDvS|]F)5vSsiExţ=8#r&ᘡĩDȈ\d cRKw*#zJ9tT :<XK*ɤwoJarExfKB4t@y[6OO6qDfEz]1,ʹB֒H ֱw;SpM8hGG&ƫEJި_1N`Ac2 GP)"nD&D #-aGoz%<ѡh (jF9L`fMN]eʮ"3_q7:.rRGT;}:֪a$)gPSj0j3hLư/7:D-F۶c}87uixoxG+5EekV{:_d* |a%ĉUHSR0=>u)oQCC;^u'}8H0]+ES,n?UU{ x~ʓOy_>?/>l8MrHID2VSsMX^"NۯDc558c&'K0L /C5YDqNe~ض˸nErc֋@aw*r܀0 a {RQXV-/p:MP\<=<^越a/bz?ܓvjIg3MN4:]U]STa,@OKdĹgJ8@o2k'Hr~4Z(I8!H G8HNW%1Tќ^?G(" 뭗R==9!nKErHc1FYbQ F;v?ob-ڈFalG*rEX}HAP'Hҷ$qM9(AHx!AF 26qxCdP!NZgҽ9l*(H Žڒ;̼|%D Ɖ`Pj . ֈ,ixp`ttOKBDޙ''aLA2s0(G2E<I:xsB.ȼ*d42I:<ŋu#~us{dW<2~sQ37.&lOľu74c?MՏړ@ -N*CB=i3,qjGkUտu6k Cb8hs&sM@-=X(i7=@He%ISd$&iA|i MiʏݸT{r[j顒x.Ƞ"m@Hy_I )j|s#RLyL B EM;oH$$]?4~YrXY%Ο@oHwlXiW\ΡbN}l4VX|"0]! YcVi)@kF;'ta%*xU㔸,A|@WJfVP6`ڼ3qY.[U BTR0u$$hG$0NpF]\ݗe$?# #:001w<{{B\rhGg JGIެE.:zYrY{*2lVǻXEB6;5NE#eb3aīNLd&@yz\?))H;h\ߍ5S&(w9Z,K44|<#EkqTkOtW]﮶f=.*LD6%#-tңx%>MZ'0-bB$ !)6@I<#`L8턻r\Kuz*]}%b<$$^LJ<\HGbIqܢcZW {jfѐ6 QڣPt[:GfCN ILhbB.*IH7xʹǙMVA*J'W)@9 Ѷ6jىY* 85{pMX+]o$h{KrҎl 5sÁbNW\: "HK<bdYL_Dd)VpA@A i"j<鮗 qwc&dXV0e[g#B4x╙✑3'-i{SEȢbK6}{Ⱥi!ma0o xI0&" 9cT)0ߢ5ڦ==!LgdJΆmΉO]T"DĊKٙ@qP,i Nl:6'5R.j,&tK*iOFsk6[E__0pw=͠qj@o5iX0v\fk= ;H J/,t%Rwó^;n1z"8 P޿[V!ye]VZRԾ|“qNpѓVZD2"VN-m2do9 'H*IM}J ZaG%qn*WE^k1v3ڣjm7>ƽl' ,Τ9)%@ wl42iG.y3bBA{pR A ?IEY ?|-nz#}~f ‰dŷ=ɀ,m7VyIwGHέ 2tޞߛM{FL\#a s.3\}*=#uL#]  GE|FKi3&,ۓxmF͉lG$mN$!;ߑl5O$}D~5| 01 S?tq6cl]M[I5'ոfiҞ:Z YՑ"jyKWk^dd@U_a4/vvV qHMI{+']1m]<$*YP7g# s!8!ߐ>'4k7/KwΦθW'?~>x0_>9Hhs%y{#iUI[Gzďx7OnuKRv'm;/~n-KI`5-'YݦD-!+Y򼤙&m^YAKC˴vҢ]+X`iDf?U7_nMBLϸY&0Ro6Qžl+nݷ" 㬙g|ӱFB@qNx^eCSW3\ZSA !c/!b"'9k I S2=bgj쯏W?=`}H0--VV#YmKW^[?R$+ +cU )?wW@!j-gw2ŝl1!iaI%~`{Tռl>~,?5D K\gd(ZH8@x~5w.4\h(`dc)}1Kqi4~'p!;_V>&M!s}FDͳ֧0O*Vr/tdQu!4YhdqT nXeb|Ivż7>! &ĊL:}3*8&6f5 %>~R݄}WgѨ@OĹCtWai4AY!XH _pw騋[b[%/d>. !Df~;)(Oy )r#.<]]i-*ػ-f24qlT1  jL>1qY|\䛧\|r>Ch}Ϊ=jnk?p ^C8"M#Eޑ-5@f,|Ά(Շ*(XCK*"pXR[كrq IH!6=Ocnи%G"|ڔ^kПy׏<:n:!d#[7>^.hd/}ӾP'k2MؤYy/{!ca /^wT j˚ب|MLE7Ee/I lu//j8MoGqdDt^_Y\-8!ד|$@D.ݮl`p48io^.š{_f>O)J=iwwӑ؇n-i3,1׿5'odۆ3(h>1UW蚍R$W>sngir^$W v:?_ͬ5kݰw[!$s׭dֲcUh=Ɩ9b&2} -/f;M.~dhÓ5¨LIa6PnzɗBQiG'CXt!*<0U-(qc;}*CiKe@p&Em&x!i6ٱ˭K& FCfJ9%ٕQ·BD-]R1#]TROr}S [;Zcq6xMY 6seAU9c>Xf~TTX)QӅtӚe~=WtX-sJb?U'3X7J4l+Cj%LPFxŰAVG Y%.9Vnd8? ǫjU3k%E)OD:"Ϳ%E)=}l/'O"Q_4ILAٍKK7'lWQVm0c:%UEhZ].1lcazn2ͦ_DQP/2 re%_bR~r9_7*vrv |S.Z!rV%¢EN$i^B^rX؆ z1ǡXtiK`uk&LO./!Z&p:ˏ!_B{{s1>"=b'K=}|+: :8au"N@#=Ugzy]sTv||Aec Xi.gL'—Ʃb4AUqػ< &}BIrwZ\"t%>6ES5oaPqobb,v 2w s1,jX4W->L!NUy*Gݓ KmmlTbc[O`uxOp  |T!|ik3cL_ AvG i\fs$<;uI\XAV{ˍlJsŅjЙNhwfG8>Vڇg18 O3E*dt:|X`Z)|z&V*"9U_R=Wd<)tc(߯)Y]g5>.1C( .K3g&_P9&`|8|Ldl?6o AMҪ1EzyNAtRuxyn\]q_ߍ&zk.)Eu{_rjuWݚ;*6mMq!R{QWR=oVbmyanUn.Uqsy.?W8 r[zW*8nؿ[;vmcoW]"U;gm>?Z֒Z6`!2XY]-Zcp˿˘ɲ}MV<в~!?YXV+lx)RRfb-I7p)3XɯEr^,bfbKJ'@hX><[@ ,&,]$*բk-Yv5 '1T9!(*t 0'b@񲱥-kc6VnR0h& 0Z|ђ8 CGV[4xIIWN?Yt>lf@ Vi`D~ڇŁQLLkY <ZPKoma_u` !>Z;3F\dEB n+0Z ?&s{ 6(E|<ޭLk1Yn(F!%sx]>CTl9"و5 |ݹր|/#.w0ޒx"khD?O`-9C| &8֨O8VH5uH)28 Ǿ-R9~ +#e;U6]aD6Xzqd5y n';)VKL]O@b OIAG Lmc 2;\d˽$Mu>WmCEQuabAJ;`uy-u.M>9VsWٔo RS`S#m8k;(WAXq 8@+S@+' 8U˜z+ZU;=eTtX->9U-q .AV/|\ǔ%&$]1YINJ2]:a0OWvI.O6xMY0/M$ *s5x{gsəL3{$)ՆbG(}1wt!wVf;I&Xi43غgR 6 ݩJ$)}Ta@ nS*X#r#v6*;WJ-_@q.+?DK១btMp1 1Gȩ f,M`,Lr6E} m"8_SK$_#O;V 7=xLOu-ȹ2NKLjp*: 'SasyrFrcC0 ѱ LKV:U} -:U8t[=EAV$=i[mhm"roe5jqf$i>;V0eOޞ4ccc2J1TN.7q;"sդSP) 0v3-)-ٕAg"pZ: "ka+n!e߮lɹL V3Os\ဝ+A= 2䣔AzG\ ` \vc"Kj61O Px"3Pc /' PW*3GX liWv-6W&)cX |]O;C%8@*Z1%8Gk@5^NtY"Fbi8D'+_1&1 7U^k6v읨gQ`LRx+I&s5Www` q:cdʰ H`X;"}B=-/M~C>''1R[sdJm RD3Q{)bJatdq>*Ct/GǍ-`2:u)"\**dPdvc& HwMlF@a5`+F>ΰ-q>0*s%Q)L>$ćYV\dsEGز/:ٕycZtO 2ze31cDB/eWy!A/V4cbpWaPBIpqS<(lȣ'3K?e Z?ڠ8VSZM}pnqL f2D?mzq*a[~;DY〩b𻾋-]f8dBմVs6傊zF"daeY(R+q%sor|.v\sfa:TX%;3Xl= \k>kqBbB;t@/Cԍ)Ga[ r=nl-w/38ѮI*/=2!j\FW+[3=`BZWX Zd>t*Uǖ\*Fu6Y3[yBPj|LcwaIuR;uݷ㺾|47ߍeys=.EinE% 1zY\+͕߬VͭW_겼cazyU1wOw)Ǽn@6 |lk'Z|VZpsqL5 څB}>u)^v~,󿴝} 3+m𢛲Pz_Sp2auQAP*tLnIXA6L7 8UgKdT)*7>p{Pgi-b)>U6IXabPde Ӽ8Ģ8GɄnb'G ֤Mcv4?>HC78NE@UMc8>`TvZ:}O wmm#K~ٗPbl^XǙ'NLh׼ /5~IʖMD19.v}UKw5b٣$i2'=%2VIE; jSdd;͗̏U r,\cՅG$0V(J E^D=66!*J)b.yGk=hb_env,TGSmXr&Bd~j%"O`R<***@z;V3X&(x?XUa]/Ǭ d_<i~|?yuU!Gc~tEdz?bI5N!pG)7e?<'a]U@tJʰ^蔔^؟a^h2415H` 6^z`Ra,(qB6 ?ygY\+4*CxM1y x Ln`M Ikȩa|][EUsAYQ\B[Ι*p {E?+|ߗ/[[CA1b<ox2ow'ݐȄt=]@%3 0 His%7":dm{ /IOSJ9Kׇ$:Ã[ |(pQ4 ͵1 > Gh|4]k_j-x*7uBvWg9? zuۗwiҩ{m?Ehf_E5jAV#<1`כGֿe؆؎t͒x$x/>``yc|? :)/DSkT?8%(ZFc$6a /E&*ѨL/'X]5Xr5A>ÖIaܰA=^DJ2 ʨUD%_L,:x;f@aI0UTTzyZ^dmTQg$f& & t4-jsbs epB3€x9jg8Yd~DAu)Ɓ'#Eq9j}1Px.8 X<s0a%/6y)nD,C/%& 5fݲQ"4<@ "_tELJWߗQ9YPbR.:񉉔*bj)&]rA kR"hKb|L's!&˪׳{˳U*gSg)ϷIJUlƛTg0+& [:VR'[RXu%G)'aaVDO'Fq,X$[@ $K xh;K?Vh]G`\4]4(a pP0RL'g/X˿yG oxR,YfZOgg 䥁/$1?3\ O+E SJ<2qr%=A#:&7"fqm,weBn=Gs;sw܅Im\ԮI&3#Gricogđ|ɛ` ){ԄI\}Mo)K`':އ,lQJp'^O4q.#5RPZpNK y4b 8ۑt,]'Y]j*\48qԒ6tyqE$8iʮD] eWG!K, hD/qHF]SK`=>:ٺ ^㈛s D# vAh> :O"ޜ>>E*̮xVgL~28(vRVk43IT!g84b4]eyT/n[]E8ڞbO&w|Ծ q@uEhd5(#0_`8jAkѲ>ҊFZ(xݗX&gaYsT+ ~*?!p>ZJJݗ10ʬp|rEM/ 9E.a7:`G2KWz8#ooBRP#rYP+M / ols N_]dYOʕ7̐=5e1g7c!U]ւm}AGobf8zN6n\ 8ıT9 Q"-h|vmKHq~Y[5VUlh\ȿP5 p׋,(mdbϱ^e  ȶOSe<-غȁt-p㩴l%k.+ދ?ه< XG\c$}xհLBi,]/1ɛڋ[kd}w0^,J\S'\fBhU2jxGg\/ ٷ˳*=CcQv* r\UbGHC֢EV*Ciz:I<0QD!."jX=(,פlѻJ9w&M֚G{ @vBST\x*Ϫ-WBҳX N\U|E %QԲ_ +M5?J] NS#8^O^ ET|jb[d̯ YZKWS_Z[UMo:hۼTd|U -tjeXj2p)ݗp]qtM#򡖈$X@LZĊ=GEIO?UL*ӭ^YU!{ Xj^jdH #ΛTv?[ iQ< ~M DrDZ,4֝v6nkPZh"V=Ȇq].}sʉDe]OvR!lm g' I([6\jx 'y6d抰\*ǫݻnm'r,Uz/aW@$o.VW2 Ad:@{*i"FZ1vӒ"XeJ<J! zjrx{׃,9#eyd9<ɻ8P0ُpsl%B)C /j5cTO7>>).Tm K@SWM)1xƔ ݇hKEbz[rٹc_K<=~>>#?ˈ CPuV6jkLKKEwT-О@3"C["$8]FwἎ!w62&g|U{|Q;˳GM=H󵫻6,9+LBGZGM10ʘޣܝ▱6pf%}()C/H 5dR%q(})#C0H @)ALx~9 }HS1PkO@ D ;S#xG25,&P1Nml#j8 qbpm4& ߽>5!˰7()j"fw6]~p#{pS驉ث_NMj0ߩh=+j:!̇33vCw3] 2ƛZ닎k?r9{33_j!3݇bitNPP y"vj:eJu@9[;RDftwG)bC6Q$^|Y27gw+Eʳ2Ȼ <>ӏgWnE}k-ʘ۩M 8mq 6b@i,:,!%noowcP42;-5˺A,=xy`^G(뀅;/L= %-Ҙ}Z!S2@x/~L(4 q-E=]y$ƙDgEuȢ[Ϝn8nuO XHtmkdZ zs"@lP q @ؕJ({`pCQDznf۶:HNpK_P0rp)A S93;Ҭ=:8tfaD qn< 0k0n<)p`84v& +!>_;B, ?@Xҷ % (5,zi5`cR,Q D4Xnuh8P8 orY\|Ոèe Qk'jؤ]42v(;30꽬 Љ F5m#L'Tt{N4A1b._A4Ӱ´mCdL(= c6&Gh Ę<,tf@t̺øuU3ZnjsPtqmVیmx5<~i2y{V'͌#.~!ݏYЙ ռ!4[#< #nt_ӐYZ&Z1K\.е]XN#`: mҘNV)L{ݓKfӎY^ؾHc6vku` th?>]Jwb9(2ۿ͘فҴuЮwtx"|W]WؼڼKޅ;)˶#0c|>XCБ88S~*'5sz0|*Go^߼7k<4 jZm q)-!i7; 4_UQhVٓװ~;PMȌps`EOr IzӿI"[m,y& LꘌUIyx$v p`h$mlj=6On}i"PKO8^q"(CJYVb0OEKCX \*pU~볱J.v,к a( ֭+Y-n=~RCpm9+ >Aʃ )YCLY\j; sKc Feu=9j苾hb@=7CO.~K(^3?`_!r`]O*)F^:;Op2{+vܧt\wЋaZ،. ]ߨSz~|(ὼ kӇoN[$$W0+C2JOЮqj}s\$\pykoPg>eQ?!sP=<ˍ8əp ](̴Ӂm\76#oHu¼ it2m (&Ӑ&r>h 4(mǪe_uG,È}"4}(G:Z賀CUnA)u!!YͲEh-a˲Dis zøSZf݆+E>-hyP{\2tʂ.m=QpDi ,f>ToVcm}_8<{ZP{dQm0]D/fg@gu`Et{nPWIVpJC/ A"΂pȚBP8G~x6qB`xgų0XjbQoC\ =h*4hֱC0}OO'qD!u4_hOOn+ }78!c!mh57  s-͋VrV_f2n4#}zBM7Ln%utm"GkEqvm <_B*I!kHq˴HT攕vhWBr+N=j9{({_}ҵdeyZ~PDSYY>*2ψ^`f;3 ;&%s a$5rN7Wr6FTI.8%? IY}LTu\ZhcOS~b'ϋ*M $hf("[Zi^Gue7D;&i,&ăuCS cfq?xJ(C0d Mwpo4 E>"UV_Zz(~ ӫ$ '45im)Ǐg86$ޢF ^e>ZV_DcT.է0 o=d@-<_h6:VEa8Ld) Y[_W:<oaQ)փbI4ߖ5|6=,'3(S:u:Nhd"Zܲz0/@~3Nu17K˧zQ2 ίfB"}n &IP(])\ Y~ƣ(K95`V!J[*v [2F@m!V3{F#AOR5E%ᝏ?Iag)S0{]+Ȟ#uz фLǪc*ITq#Mw>\4qy>0÷g™r㊰U}L, XsNuSUx,9I3. ')l2ZjbJI ?gd/OprdMW)Qιf=Pka(]CG4TWeI>b" Ӿq4b'I|gi5lTlQ,(ϐ+|妜eZ"8Z:+3'Bf"qtx5PWV-}8Bcz"X!e/G]ZAo,(HPx%FBO x@> ԧo_oصGS$G\ _[+J34^^ y1p!}&B0H74Z5Ktt=K:PLlu )٢]'aEc b=QQ=؃_e$J\ɱ`˵δjŖȑ'`gx>cыK}śXޣEwU[><RQ}<|AgR:J-Ztk:V,5( uoȥlAb?|,$a#ikgG8y2@gMLWMJ$ၟ_f2n;- p{ {켕<kĜ?ҖE #6:}(mFy.k\o]m[&2SuQƒo1Q +)ƲM%EQbpkd>&C$cb_#bKt*RdoѾ iʹv*8_-\ՈQf0>J񿍣;? J[DyD2 M
YOO~m2yoλE80 CJ񻄍79c | 37C|u3#kW>>1]muΚ|}Ugkfj:T1w& rPg7mr k9۰7n[xB{G|ʼL//.3>3yjk(5_ O_W3a;chs7ڃaaX}Bۿɭ>7i,~4eTa]RpPJ TBRXS}>næL#6RCs Nע Qb*1mqNRh{ކ>QXݫ)l[ -;#04%"H%(.E,k6}׷OKpQDl.tQEEMdh z32Y8,- QD #Q9#҉3>ٵYOu+WPzP9p6h]"ib l Rfz(xXLd3pRtF#0I$ T|Qy{! \Q,u$کJ  T 8FεP)w#+5dp|Dmw}{a.NGJ͙)7cՎ}KBxăpQSW{sVzٴRؾQ^.W/'G1nq)Pܔ7ˡFd馏b[td%F1sh臝KR14:$1 $3tSMDZ̥XGlx m@Hl+&j r *2x`E&N"~=ptL, 4>npF)3HuH֔ bRR#4X:ǭO]p\|5¤]*]sl0*#48,. Q{ I()ReHtw_cj̹G}twTzQVBрCڃΆ\KK'8m޾tQ%9XX ૮x DB̒4b "]p-p{&:R|pa(.Vr"X:>?].8Q[6ܓlRpy h) sɲ ja+E22%rVr&XrP$D(`LBA*"+!QCԿ]Va 6>~Y}IԔ4kL >4$Qګ;xB նӥ'רX0YȁU4fiKr:Ţr#9Oԥڥܚ,xko(E@_35CkRࡅЙ)qY^D*瘤]K}/@e仝 pJCblOLsQ:ޭ9G خ_^5W@5mq&bLѤ_ WZ+*zA&j2jb-` MZU=1~!yK/ 6 H.b̠#ݪU!YcNR_(f1ocx w=h˪Q$;ckFjAaT '5GD2Y{E>V4̼?CpgjTS)fK>Ho}pn肣1cqLĠ`ՂjوP23} *u,bqebo`͢Z<;78=.& Z@~ 3z;z{otȃc  }BX‘ C&RkZs]pc3bG¨yqت.3bR +gĢ^fQ#4U.}sҝ9`Insxpg'OiFh{- z8SDm㧠TAV[qHGb[頝i3jL.8~TZIi~YpV#R:bM_8xgyHEZYo\BMdKC]p\s05X siz**VRO+L:7Y`Ib~0D2 kuTdmS-f:;96ۈ)EoI֫{+@#R#) ,p@a:vqMʖ<3ldSk$&&"SKH.8:;ÉLw '`P.p.ED| 9⹭]J $d%s8$O*^_7_ݢ\ڟ@Ql̢2YM+9W[]nwa{YKc-*)+8 JNǢ ^ =@siOe?0vnˇ뇋nS}O*kM9 IUdńg[5q䧋m )f=K)Y;׎%\oʹG%]O7֡kS[F=23ZPlcM[oKof%l WH/JIkvlKZuS 6c: ~˹]^ZL7Dt?lR ^($n}R/]pqaet* b]7i_|-J>r~qĖ-NKOۺg6-*+r05sc҆Uҋ Lx{c B۔exR-jӑ:>L*#G\O]PTҒ:])BRKֆ()Yݕ/?Vپ\R/JT7.<\@w_]p~H :v!ݢ&NH>ً^j삣b/W5EQ1 ZG DϢ4MDUjevWOۛ.8E<ױyH9+R?O٥.8~h-+_;ܗ32bu<U>^Ijk,cJ{_)dFe/0ғt`؅h8wY%Y%nI.-U"[.$S9a|Rv9gVnh7iiٟaAmg -@|:.wr#KǾ!X}|.ܑϡn: >KQh#-zCY=ۀgVޠKFn/8oId2F%$d* 7z%3mz-ufQX.[6qY{|ַʃGl/mvURMxRm#LBܘ.54 $OݸMy2+䴜W* ƻZ]sf=?==tY8v:4OZ""ܤ\R~Kb]bx=a-ռhj0t*$2u?|$~ j9jxi?GN*'Čޜ:!2S(: %a8*D5{&0 !? ~x_'.VEi)ao'W)Bcϳ Iٻ`8VrЙsǥe1ku_,~7ƙe`8Bga=e{n_"GY8Zevw2'0xhλODTڕDD1 ZeUKgs6I0Sp (]*ӯ`QoJL{:{v{{.^: ٟY;[2">A;qrc>=LǷ?Dޛ,<>aOIGI-^ץh]8 ҽ*wZ]OPfߏ`wK(Y p|>|7=g {f\3%i"?d|2>hc;>9G+!_@ u$(DhYxDWzo㎼]Q,W@\<1q%oGh%[`~6c{q&5c̓Ҏ^E3Zq2L8K NgdnbJUO,QVzfY^$WC<bu2]qDAb|멋D;Va4gSc>v @=PV  (lˈ|_#0>%ΌMFa?nY<e20 *6tf֥>vGǭȹUe VIPv¼~3zFa߶Nê$ޛ+UqUgѰNq>803w⌃rn96Wii!9U_ˏܚytxe.Zq"\D[8g#J+Io',zvWzUcH8UzS,5q]bE'?s~L9?RF,9|fڀ[#NR: _+S;[& 2d86zR:bJ;,޿J!B~8Zd|Q^(dzPNՋFTv]aXTnDPtHo'?ݛ ; 3Ucj)>vex^Nσ?b|f"4H^yFriJM6)n֢,{w;Z`lFY# nn%ԉTaBǂvMFK+RXc/oQdFqgwJ)(p猔(B{2J oe|50 ) &T;!0 ej.7<Y]um=8$ 5;GyS|@g[=vlma5M6kGa0m%M k9 s4wZ`3bO AmMG 3 '5ֆ3pq0c|UmXeY EHL!okLj~4/\= WQ\՚j6Y/VSnr>y -h'Y<wJ%p̆KiŪTs%)h,&rm-u8i^WV6" =K_FR'p_I=Z0B5f>uڗ}|ɋFQ;/{"BKe4eٛs*P}M%`:}ueqRW0??ߖ#{@] z*C :ܜs"v||9h_ٍ־r8n~ $9us/0P헑U#~ iƿ% bv3`cP_%j׌{sxyHaH:p1!F9u[ 9rvR}:Uhuz:(o10L7i57!H*ܫ8x/+< ^ox;1cr0'A2%gm_w %P;{Q^ܩ1 ^N\7: m>HҩQil$%΂QmE2 uV^l/Qqm!"=Ao3.,fﴠEuaGołp,U-VL yQNzo=)Ҏpd8_Y-rbZ\ZZ\%UьuaLU3J7"cUK4ʁ:ZJK O92;U3X3vTK3vt~Baִ7wh^Wx*m O*Uy‚R z2^=Lu Ƹ!zR8k#iJ6扺≺?]3t9f0pހMҠ-r\:Rp1xx" b3e3))@{u;u; 2iU/ Qk~IF$@M-Ω\81r Bߒe{QljJx0J%&IԢ$zx]|}0} $:zۯzY?sՖB>OW+Bnm5[;E?z ^s>Dɴ#8yn7ΊLXՏ*LWu Ho>&miVOKIF_+3h)6Ƿ?|s6)ӈVz|j{Ssm'z䛙wo.@e᛺_5掁rduiM`zך,|kbM5ԂO b}uMzׂ Ļj녵E4ץo7Ans4\L`$\L l}SvZxø`qϸٯ\Vi ߛJ4\:hGe;6㛧֚P!֋n}fha 0]㙮o毋`uW=5U|!3m%/ޞ$f]` -`-=8scn>c !(֫1vUz-G^R\ɎJmpW)8]wϵ>nQ`_UV)Y')YSA5qQyb1TN5|DJ.j& hgR_b<+0WC۞+Tܴd _h5ث|\?m>o\jK]82z^ll1/iSf󒊐 _] bzY`KN:`kre)E8 MM, DBu^pu+_uPT5mW86FP+1+Q[F5칭Q̤Fni0q*))!iRNIJ0wbn n܃1Zl )55{-QJ1iaVc-j5ƝKv@)wH"`s# ރqJ@vT*刵HT<; ;6: 14"Vk5 #A4[X^9@J)_Woq%mDIgmqRoL "9o)>9NtT2LRB{ d>MKV,:uQC$ǥ`R89Mnblqwwk+.4ӆ4g’єÿ#QیH$Ziln'|ǹ[ڈ6.H 7!T6uIDU{  \6D$0bKR򌃷a Tt>ޔst$;6)5mR2wk}!Z0⚠vyvҚD6it{GGHu@]*y5(xùX [ *l+ ɟ ڽߒ2QN[N#." Z.NC5KqTWUFZxDYdP R KL 1ȹxp H!FHQa %/x$J )h5FcA|j)8$iѨ9B L[s 1VXMN)aE2D2W8(mB:cnj$U[ZjBS#)m%00þea߲@˻k|ʈWi@r T(A:eX*\! ØF3 NPSw2eH@\*CRrj5yÔ|SkmHe@6VU NyXcl`;q`3Y%%'Z5%K$3Flp (RgIe yYO^sŝ9K%`Y ҍ*#K\Al)ʤE+tVh\JS(UY)Iqo0(\o(H+ S"@W" UMvws ;NbHuf BYuj8C3;ۦ'M7l43q5R@|9H2ExLu8A'P73jRuv3'Mhy"8C(NM7! PYNSXVGKQ>K& R!,G64X!yCC=qo-jdܙpr"@ T(?tˏ_<||au҅2_(66\'>J\FgFIAf<۩Q [LD^iTaUʚM`S ӭt/*5f:at-T,+̵–id1#df$hDMTlTIA.D].FAD$g㢘v35|o?eB8Yae9E+v7క\{VBe J?(+eip&c Zڬi1rR2V,( ˳B̓1S)-rK;SF =2Yd)UdVܰB qZBe%Π%+3 @uGTs nYYХT5Kmi!O+(uJ+K+U醢2CRTFk(+OE1x2U3JUilnSBMD%֥ VpGO>]ғFdi/X22DSU Z5DkMQJ0EeVUx gUUӘ8XUr4+,K/ITĪ&S~"Ůa$u[_C%;58pl}:34O0!WB/ۑT\/V%*v* |yMB+.UáX%s_?30CЙe:ӕehlpvi689QC͵V)N7)!m8~?%qvhp%ʢsN/0WuJ>.iEm{8n|>3t%oK{oַElWWwiC =|ęs0k?.LĚeYEU Mo1.8V/ϫ켸>D N L]|׏3Noo 6Z'3~uMJzM9:A56;Cż,V(+7F,WBdء685EMX|4{'$&ٙ\z-NfT)'KM:[u2wC+Aά& !Z6a؍UIpk !(m} mDE&e{`j7r)" owe8%D]k"`ZXޙEpe䆥Pm/d' dҷb-B Lmt)CD7Id*!לu ~su[y:I'9wr<}NSc=+/w߽M{PzMGGP0Mżh̿p6ć ғ_~>?Z5V=91 ZwzKX\DgF ÈU}4"`JG8$VFf5!Uuu1Ö2 2Fwhy!`+3լ+df+ Pq Ҩ]h"iz:I+2<9) GvIk2ҫ˃3' N":>u38KL1*xR=^/9=9.WwT3ͻtpwžqӑ4*m%pPe&l`^,a~9$YӻW/3cb_ҕ8b'7!^=ə+7{ۍL^~}w X̠ ~ʨ_^^ϣ٧Ff<'8O/FMPiТ!?-Mg|t yuˆ 4=٫?M.?΃$=Cu@C,7{(!ߓflӖ7^>s'mI9 u17$08L)uڜ(&&|Ha0wf_!ї{`3(FO'zrYҳ+97اzY z y?g@?Q?9>L F8}g>aՖ*iYRnRŇ\p85rGcC)&ߚGiFV>L4{vkPa=4 ''SԠy}]_w4jm߼M߯fwv[{Y‡PM6ؔƅRuѧ[֩w4>v]AqGV/uWWi nPBἑd7n+5M\o= Mv:f_[j`K0 LaEt:`'bȅ`Lfuh6ȧp68[IڣQjp'tk첶AM2A`%O>l4›Uծ~^GogU73'FU$ٗ%y+j!w:oN;J SF> 0at0i@Pq}ie~DOdG)V $#7&•q{.tAjР] (=Îށ":P@V~7Vh;{:#\;[纶ucnvc eoyfI>g"/o›3b:+ۥզ4QzG&PƻLu$:RFL nMG1ZIj?/Y-x^Lik9b"{9vK9|nm4:}\*5z}ԟ7F>jNp I7 X:\4x\5P@1ƭ|Ε!i;5m+鶼LB熊380 W6Qm(I.*pzv*Yݳ,ay__/.h4Ί1mM*:2BS"Q]w~n78馜)oIg 5CƘbEgy5e tONѐ~a烽)qN!cUY,ǟ!5[˫ҕܕP i-D3f nxS<ϜIs˞efyYl:}=Fusĵ|zrO|$5Qas\HѨ' Ϗ~(٣U`ϺsJZ#y(jݕأun/&C[/x1/xډFm^_u%*x1fhn*oѦ{@mYaGỦP,/G7VD=;/J@FYFm#6PŎ7丮'H4WL1;^ Øvb:O1qUX  "ʋt3ZFt F=`:"\" @vF1伋{h-xsDWt\pQ.8ڬrYn"YuLѭ: 0,U Dj{bRM!(rqVBN#Jܳg0'A=r1Hԙ^T}Њy`)MEE0}]9͕l}mDFrI0*[Mlu\jИn5j( ׍l}T.}w_ݢoXo۷>,NH$I^+ 2rj=nWpG pH"jͤv|EdM"\e\R6Vb@JDmZf${)UkRE؞ @h:A"6͔%?pDZ-AN9h{^U͜=XV78P L0ѵAM-Tkp3Щe.Z,Z”\`:N@UZ6LYs NQ_U=D-׻:XݲNjj6Md}Vh-S6(0ODb瘐-HhAHG}dWZoyZg}:YT*^TA#kMUYy9 AewZfIh IT,S-i}(\[FBk$@;|(hk`z >TTԚKקP}(Z%l(S\.Jym!;R\jlVmg)I]o1n@>6ԯ@-xaBk&Pq=-2Mӛ;,|6/(LNSP4a &8`:"`gu$E>'U;,XhaweVfH%!²[2[[<*-)LEQ;>J͡1R ׿-Bs'qr•VmD$'󉝬X`#";4ubIhmG̻ ^a ׶ha*bY 5qHFuM OFOtOz6=@8@m{ }j% F^wk = P'-t̻^s(wo݈t(dz_̎<%?დzrYҳ;r e% t+I9NޤʕI"͌I忧A} sjIj<< (5?})iJguU|oχCSC0;~m#XKw`m⺤Z `?Pp\hޱ;ƦD|vLV}].X*.gF:ӳfL])sRַ]CCCL'b|`>-X;>Rͳ {Ifa^k*nZžh]ocڝ{l6ھq硼&'!BbHu]^W1]/w0ׯtV7nc޶&-bh kp5oЏk#tJwzYAtz%|5N L$Q˿c.r|oxo ν}].Gz׌o% ?M۷k^2Ld!O;fO}mѓ~B/}ZB՞Ax/fw;سeBx'NGhɹR|'gf%Hk +( ϘjI&zahO}3ujcczx]T2vFI P ϴgѭZ `/^z$NoIDS/CAʑ"$S$7][ ^u9zOԳq)H)rD6=;9r2xL/:! b\R=g!ƦO-Y?}/4 Ok_|WnW%CQ|d`tKHH \3FѸjݙjwkRYW+">wkgp\\-}m~K'b}w|~]t³m ?v&[̾oMhGm#Yr]۞ՏTV"\ MXj H1"²={_=:^;c]-4b=_^p+Rar谏6̈́t=HOt~̥ _MiVP0&NB3fj1H1 T*=_.//79^mۻʹ|}Eܤ z֌G]n }0 e:tҋQ/_o+ hr)4c KHG!KbiF+B0fp\&-W×M76Ynhq`QO<]*Q-O0v[86NEbTqr]`FmiOR R=?]s|UK 0QBϡrrbl| qi4hϿh׳q6>C idЩ%cȮqbLȑ帛 pͺRVZ5_O{T~9tn5Tȃ1,A=6 2(~iЬXV8OP4R# YZ~}jZo[˩kBsOF8',+H1Pk3̫0zij3!aG; Gγ2GK|GPkԈj,rDBLyIbREsZ-D ;j!i![BhLKIL ,)#J)V#8HT9vttv cܨ}C> #C;mذ"kK޻|uB1@/zGct(DSwi,-)bܝss-409ͱ.[#ͱD8@Rb@Luj敩6D;S3{,puUa5W55BT7f}m=1[TJ9)AyGϷ=2:s%: Vb#cJz֊MFkE$,  R K+',E,9(KqYHncIT]Kƨ% ,ɯr*Kܱq w_.9WpWq^zϧT59tB+$2ڏ\;N@xlQa1VqbW L p$@Es!rA\4*^\<rǢܛ/d'm2' +8rbžTN*]E7w.b p}A"(rr"jg/*e|c?X*쿲 @ƞ.~eu5˯1zzF t(-"E:Z Q XHe$nۦ類2ROr0_oE~jLeho㘿lfdK <жQ0Qp9 p0 N_2q9H1-)|ESQB3TK0>d c' D^3H RD"i!H;>9y[>$xvC4Dnت~* gxhÕ P:C%/YAO㇌ZYV;*yN#iOб?[2su-ݦc|M!34`8`!g!e1H1&x&ivMRg,Y,xe!Q•d!<Q|++cU\(3%\jٿѻH3ԴbbApDc8Y1u^5g^x-M*@ge8Ŝ1țb91@t&B_R102Ï h l/[M,jPxHfP z!$ Vm!Z ːpFr$&W R*m 9HXY#Gl H 3~7/~ -S$@>=[xn{oϊP<H 0QcH8CAʑKzq*$b␒FZL[ؓb#ip5ɠ\%ꭻ76ȀyC2-Dv0ydZ19eXbltѽS)F",,)+Z"+/q%{a+&'5z!$ s A+hP^tb ڈDtE5=z9H9Nߞ;2+Zy q1.%,)*X'YVA%;UlV=)1'RrE`VnR6_=ONT]DtR RǒJ~,ߚ6cmzKo5 du?Ⱥ酭-+VbvnD:l`vx;[68tݽpw7绻Mxm\Ø݃}1Y^^o4JRF\v=/I5!lǒ]"ؒL^^KRWTtѸ6L[zveL2h ORw/L/% L[PZ<ŠGtueZǾBY5e#WgϤ zFIk3 @9ltB677‚gYCa/\áwˡKB r%DonEܔP$$fV*䞑 :*%Npgg 9UV#w}>-l|G֟0a9~ypő+8wS׳vychf.HqN/xRnK\?p-'E`RunPWQ):L)m-U3Zc_'G6/B"צw4WWՕոn(E贤c܊<$fkb#"xcۥh-lb{2z1@c>d#+YYUVj$73Ŕ#$Gv߸Y7Ye+ʼn\i7^ڃOsrdN/-Ð:Zܤr 0aSmR Ne}_e4ZA]%$-W,L_ ozkE%vѥ.mtץNb) ,R$F9(T=GwYѱ}Z:vIF4R-#/rDc4oTp~_Ŏ`K2u{J%?ExtBwОJHaQPBAVʣyHPJFٗ[g<#[˺5t1+Ia5\LyfrOIC0H@2E2o /ێnčėZ:xYzV˶EZW+&{4/'4:\8YoQ'T[?oG(cՍt9Q:@>Š̽ w#z3QSq)(S֭!%)`,ncF!Xu6<"BfY%mVIU{JJ3RzC2Ɍ`J.d!ZѬ6*iJNF[fwBF'X }F6:ѱ[DzIn&i ؝.%~eG)COowarcoQ9ÔOSx^_|h]ϲA~ ؤ38fZm[ۋow{>햛&.w$#m1My#@ޖϿF B(8M'[.u`3 ] .] 7/1c@~ٮLWY0W7訐@jY6VZ'|Ht:=qƾ?L5Y4ؖ|3p㎻G4Dc_x?󶞴Oht9蚡 yn.z%R?hKL|&?ϊGnp5OsG/? /hyZUSu"qxOtݜ bFIk87aO8}o1S`"ϘDUlaїMD30)XdOu뫤<ゑ~Xw?#aOQFiQ&)`Nٙ)\NKu !10nBX6auN([d]bqF/K[5`RU 0̴Uf祭hok;l77-f`1M39ڴaL4_v>N1Lk~x̮S&'sپyEʏ0=Z͆fOƉ#nowAYLSn,F+ZvLbXꕤ镟ثMiƤTv܊+3g^ٴ^+>AcK 9+ YuB1rXL\ ghjIQqJ*\H%6Y?\,[e,[ehR5)ثM!֥k,CX945LD̉z0JPH3jǿ %+ޑakC!;fHjUڐ<-T !߅ah;Lmy`9)JUm8wpgsɜ(/c\9m)O},hM@Yk}o[goiQ`YSۯ%qYh/nro.>^%d7]ZoƁCc/~4Wo6UQE8OG)x -ZFN )W#EWА5 ^v`:`r1KE>tJli%zPLjLέ>UMP \Ϊ+/Qj+rM\[ \M.喝0LC7*vzk%NGiQ}[bsb]RHQ Xyb<ڎU`M*{ӷ:q;Z'N*>#0Z 3.c垥u֤ TŇ 1 0v0opj2 xhMI]ک kt\_rc7O<,/Ԋ[ĿWjK8ؒ/{kܝUrcLY%j.,bWg]n{Wݼ-/{m=e$VYW\T}ٛ?ƎQu5-OVs{d#PGHo_a6G#̲O!5?g6)4xs_C*ǿOqg5除 >tG!1z S'yL2;I=˨ʬPdf\\BGUN8ew.M 4Cb8'8Cν2 t4@C, XhNVSd1$NrJd*Agfkuϴe5o3MaBrΉcLje垳BOEW.?=5M%㨤ds EUС A,ڜYj}|=V1WJ|:;h50#m!jCh%a2T[]P^xY̔Θڋ5Z.:r> !RQHq I(%ZxVT JJ- T@Ha-$\ݝ|$:3`LQx34f!sea(|E'ENE9GZc'@8Όշz97*SfyQwڐ$\0obAc4 4WZ ^%EatB;9A3bBB ^"'@8`uXq$i_0ˍ*y09#*!r{;&RcP]U /hrx@nr#/0z{<p:xԊ;p,E Ekm-`'sܕSc "2.V9„e &XJd`:Z JSAe*b<|A%#`)7# `R"`&@$80&99Ϲ=!)šΨVÖiNs ^/u L+h#;6B͠j`o|ͩ B0`(SP|kLp&#%RI9n,EGCRl ;)m&b焟 ]FkX5{QR1bɶ yթJ_J-f,E40PZ- 6 jel $z a!ی꼵 F Lo ϘA ~RxŬQ@r|\1FL:D!DDO3vL{k=+hږ ;[.Di> j0A+w G\‘ `˦ dUumi>j\ myLA-ENp`CEE `ʃZRF<@HE&rZ!d^e(& L-xx 0|IEPV/Vgx$ ې, ѧs2YЩ ֏DE} 3r0ۊjLj&XZLp@ԍ7P7ӌJD3 58)(ا_ENi6wZ(ΣDjсaKDm2&PYq 4#BpP)Q^"`(6%TQer#Cks"EY.j:#S :(B , R!!K=x""0fA='veÆQ|]oDRe 4Iv#7H.za*@Q {P(eQFRSdPGq125gf>9X$XD Q0IB 58v*pc35~iPc- B&TߙT'#)ɐ+*-L欭O֢~v):9?VKeנ|+Mx57SA`1|+ƨ:A+@(\XJVyˤ J '΀|" =ʨޟ&pJnÄ gk|T0 =FM4(T %!.IR d~;DE!܁/ܾ?D{ٴo{;=R;?ø /n!OD]ovjqkP`9^^vywnXIloY描㯹՟w|ܙ?<6yl woo*7מ )YYy QBPߑ pFx=QOP!<>{衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衆j衾F=*beGz(FFp݃ehP@IB =WU;C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C}z(JeP ~PkB7z(CwC5ꡜJ =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =C =ץP`YS:73N"Xb|5Z`}d1[=$ؼi¯7[xv}०2mm  X|׭/(U'`Q:K^)::R/f`6hVN*i^y Rp\UUgBNJIpp/a:Ӌ͒ANHNZM z1 Eq3>r=,"uV)l/`5`Iq8Qz XfypNYwB>j`9a׸DzU_֗(Ulb~T50=Of./\^ 7ˋy L%]LE6Gf2/kqZn8́`͙Ф],M~{ '{czfoPO}J~ʣm\8]u~W\onf} c_"Wӣ/Uz?./&ͯ3wW~FwM.9vas;]j4 L~v-&؛ ɪM?+g e3/h ia/p>j=cs̔L=Ϙ3q\uGQo>G[ i6-еٚf}6s{-fW4ܪi{npOSE fK[rMNVym)ddFDbSux{ߗJ1UBE,Me\L\HjNs'g{Jgyiy~9rvOs0GokH%i]2RV첍/kPk,T2&#H*פnEƖb1dC,Z !hz!ջkP9[j͘\5#j4:I(*)V!ZIR k̘͂apuʪ޻)C)Fa]q X1zEeeq %%rxva7h߀5**R>Nx! gObUFҚw-"9g& L66 .X>.2m;~&gF]h5u W纹>}dᝏYl4{:` Q2hOBUO2X Y*K'jjdU9挑s"( f[p%ѬEkAޙW@xnFe% MưmpBjqu$~%蹉mPZ>aF6eD5(%BB>i=2`Ҧ@JfŸzѼ'Q\2c I&l.-6JRscԈ%gaEVApk'Ͳ}eSH! Qc(+٥$ِ"L%Dtȗ Ѹ j5)mzmΜFb[Q`Q nxZSΡmÏuqx(5&JR-,jl72x!OZF %´[M`lO8DY̸0o?!؇GY X`{!f~Yω2,%3A9NrmV|g<ı\b9<{Kp' 3[e=ac֡RRP",dD@]1}i%Eׂ (hd;|.bk'-`΍U"\ "U2VIQ22A.C&T2 t8h[pB(AJGl4)'L;j"$XmRa vmZLvrS57.9&$ f AX&- R A;Z @ oZQo.gN7$֔G9 h?WNP'sU'X2 /F1(ccP.i_6+m=AuD! 98.h /4%۸A׀R")e`Exs@ns1Bv9*չcAXtt ˈ*$/ lwp'm# ㎥V A}K<NI/ʀGaz17î'̔HZeJ!2@?A. jX!o96 a4L$馢|L]5,j %bl#'r(2Mr?Htd1s@f$~|D+Ũ}$kF OU >tae҂3^\M.6O8%7aBFГ*Ai9 JIJoCIb [. JCiE4fc)[h\*CC,KN1)4k&٤&(btIJ. #f J# 9ϵuԟXDq!Fi cVe[~ދv+foX}''VtH'vPGU?>Әy/?x;qO&:A.'#CTfٯޜwrrK)˼1F9r~ӛ]t''g^ֿܐ7zo\wS?Mi~??wZ'3@\;/k^^aUcq~NFR#1@jQ: ?QG(0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 ΋5g<'QS'77èh QguQguQguQguQguQguQguQguQguQguQguQguQguQguQguQguQguQguQguQguQguQguQgu^Q*Ȩӳ1tŨtNG<:FFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFaFay1Fw{yi˾_f{f/,_se.ς+vzO.67r׮rnlw6}B~* X=LzkfVK(b9nȎ`TJ XmT`5Ȗƀ,ع,,wf&`qed[ن*|\#5[#+Wx_>\\DP~g]ue?x볏z6;>'^]zN=y\ Xj`y&`=\ ;uG { =NGkvkPlwhgpPݍX29 Xd.s`P% 3F2!3Y:g|n`r\nwĭ XXY.:}XrfcQ3=qV\?t&` ^,NW޻:=LZo%f[r'/QZۃ?cp=jo{5kiWụcS7MoݫO@?ܣQeTBSC3Zg\֤ht$ʫeVsӷV&/0W?&qWzsζISߟr<\𶿜-w\^gτO-HS`6M:O|\?tm-TEo?zWSMqS%8  A#SO7Tm`w:L`;X\€X^:yvޕ?2"|ʽ`&92~rYm~eӫ}n{Ay@P6v־ׇ{O}v;ew|WlķY,Hl?_,_G {7QxM܋,G;@OoT' 4FHE3ʇ~~°L×8HO?nrsl<4 nz7GAѐ4ZҪvbKNz?:7'Q{Ӥ޿js#}[4⃟a):ӢJ\ T%U )^ؒՅ3KI>iVVŢP 1br_Uq1?i,ν]'>h?0YK!_*{C wQIjɽea@&jM $aX+"Ԉ$:Zpl*7#h .ߴmEǖb1pr$Z ,yyqj*Zc%|cTLz{z \5ŪT seb11&3ncV{mʦ6K)SVUB)F|h`@fQvV%KoʒR/ǵ; nEVE4P&RM>l0wG x'rōƬbKk[G8+~"J?#ٴ5'&xc/YlqG Q%C$% U{{R\=oNB}VeUM)1`>iAbTQtQMwmq$W62##ovż,"$ 5OtS$u))֠;)>d7NeEFʓ)"tOV)dw"ҢOuT]T[wB !TGhOM6TuGݎ05Q/3^jM9Xv)Wu (@d+c{%t@>)p-a\ǯ:3mqKAĹU X.*C$gKgǚژ[]ʢA"&⎲ q;fuLb84iȠ ȇ5ڛQ̈KUU;f]瓎1!Eͫ"Ik4B mB%2>k>ݸZ015.dqb!"ka&k}-7Tg-*w%S0+AB+⛮#h`SbP2X|`BgąGwל@P$rAV fo8wvCH`YDмL}L 5dL5EvTn~uXtu~U& :QI5%PU;G@6tF_ #Hb"җRmSk/]@,M%C.:Tdeʘ}r6FBKB>ea9hP)_C߅^:.'t׀R"T4e`Exkm3m>%t MIJ w ԁ>鵈 *-fEV!m!U9M(W,Z{fH25;{fe6Bg$2SkU(ͮD1"GU$CGRE*{* !뮒BiYfJ];Qb5\,ؚY<4#liِFj@gVo-ZR譪#o}Ȩ *50 f}`T Z)mm`=q;7iޭtlsnr$Y T 2f%F63IҢG*wPIlQfqqֶaXS!y֚BKBs4JVO ]z31mWZs._ 4#D pлaRDluPck* F mMr%5n *X1Z`ѐ2. SU"=ke U4V+Y|ow`E]!6b]3Fp-{c~M *aS'D }4#p.W=@UlU?dNԫQ3 QA,eJ C\q%n d~?DE2XoCn՘M5br"&b9\PHlR.oBGCLZN]hEWKl$Ti_&]~B:+kp!GUҋ`A_/_ZܼlKwkZNA)OS+4"R"ʷK~Aszwށ([Sr<2DsN{-=k1/bdSvžz[һwo-lޞO*S7D>v|x}fz&RK ͯ3]n>lJ>\|ײ7}:^YHHPG8PG܅:.OS;n uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uPg uP uɍ$!44PpY #ZP(NHdB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)ԙB)yBvHBY]0P%0NQ5LLLLLLLLLLLLLLLLLLLLLLLγ|+=ɋ.d%n{z/~Y<>.)z׶u7dw/{xݙ7\v;;>al'LA"3Qʳaڣed[ `'7J̢9N V8Jz dM] (W"?p~P ;v_~>*6 y J;*x RcA@ `M GYc(`]Ef; )5XY|ܢ@0PMtVS䙠 =r< XwĬCGXu<(+QHVTѐAZARtܻ̏k4 X(Lw1#( `}  V 6HK/R `ч`ka0`%AZE:N67}&`2я6zIlTֆRDs XR$^`Ǟ`%0OE#*xk(()Qbkَ%5.V`r s!]])_݇vr+'=Im*o^&=l7$PqpÏ 1Âq j/( `6|ɇ%j (aȃL:Ynl@6(L46a@"| `ecK ֐QD;JXb::b `=YV %P P&O H5‘2c?;5yp<ƍ2hF` юѐw- n& V&b,E>`q֎,; ` ڏ2c~=ۑS7zFG(Ѫ#?}܇o<1hIYOU6Miw7??~T?{[gP$REE1jLة6d('҉o @$;m]nkcpE{pr?}ď/"7gz ΖM+oOEܜ]l𶿝mߜ}z?O_f|{las  x?-O[}vɶTU OŤ7H};n4D﹜[@[lowtG !?Κ _xF_x:jO`\$ٜ=F_๾h*- ò0g !|[yA#CY)~EBDubKġ!z.8i)q8]H1|b }856_)']=pW9&k`gN;Z]m _?~MeNuhF՘Q*6d}ʹRdV5wK[[LzU-Wjeߺ%F`h?bXꨬ}֖[-E'>KݺV|-B:VqL#j QS)mvٸCLes~j)zLnsn[WxlMsjJ1'Zj; Pdeb_hVr2P5@pQ-^rJ h`QogՅSWq:ɤxCڰ],8Ye *CQ"sLf!BȥBcVYjO٢pI%X3WLd3<#KƜ 9s,/H)O͹xx\eVU˝TR8j(%%銫Euz9%x:%[vn{˥Np -ڒBGYG'&7)Z41h--r xD&U #BsmH_!if1~1`,v3~ ڂ-Q)O}됖e+H㊑X%=}:S !kM]^ajuc+${͋z#OAQ ׄOQdO]X)iDwE¥͟~}륐]!(dGٞD }m.5ˎaJ$|*/V[k\4{mEVdb {'൰N_;wMA[Q&6\x_٪c5mB7LW* "VjVa EwF&+(\ I#ڷ ( ꛐIv4*.ӉPu5%M%$b0A+lb7|dPwVrELAAX7[QPY13mS0C3%̒2֜G(kqd}0LX;"s`L biƗ)Ā;&|B(6:ڸ ֑tGJ0 Lpf+%R92rFj0^[5#|\e 6#}UmP& J۩r{6-z=LIAPٗTnO}h+/OXH&9z--ġFt{mwԗ8$CuH*мGwU+czh2&MldPgm4XڛQ̘HW:|*ƄULl^vXNRBL&DY0w{Ǯ]-XX^ ڛ|-7P9olKo:@V*&{X\!iBe ,4THMtVk YUBwU&E,VPVp*FX6i =CL iDF:̈rco^kOQQa,iJ C\q&BU0ט8@\T"nmuΦrшUri$ s9\nҪ\ ~IF isB0p$H$E yۨO<"B.8A7UQ~Ջ6Bp| 8 kvҪt k8ҏ#(fJ[ޣ8=;ZS1bA9N9'P=}ܡޯ颢j]_r}8{h_/޾51X˫6]Ne*_>NՍR|vqH|oy ~ǚe%bVg2SrIoz>3[͜fcpOn:=6bwz\^.JG>MN=fz\S=e}hE+˛/hUGPpYNC-㠱$ׄQx(RÔC<7CxPC1<PC1<PC1<PC1<PC1<PC1<PC1<PC1<PC1<PC1<PC1<PC1OC[>1 _6E2 PNzby(桘by(桘by(桘by(桘by(桘by(桘by(桘by(桘by(桘by(桘by(桘byoWCh<F= E?y *e[p?PC1<PC1<PC1<PC1<PC1<PC1<PC1<PC1<PC1<PC1<PC1<PC}k<]AgG/.h-ۈ+¬RZ}8k|?̄ŬxWm9m6J^ޞѥeF!cQGyPaKD y3XD$A"14!840DL}|"b}(w*ڏ ;2 RV8gJGCOL|~Ab VAZe9-d e-KŢQ.͖b͉Z8n/VjDR^u b>XmL<zaw 6h(+jzA-a:puXi$D2҇QFV1qƘC_E^8ʜuƸ_E^,=<2g=aƫB7m)!uۋ0XzҭQV?+Áwm/Zzv b5%ty$Â@; Xk#%ֻQJ낱vK_YCed22IFYj bSF2 艤"~N#4>>;z7[8zGFjb1[w߬?]Z6gVmO$4܈sN26gKGJJs~cN a@w4|Z}5E93k;PEQ7kI[ Z, =޻POYKu@ k}˿+lQ=V%&N*W7%VNQNwwsq݌! Mw; m漜s۷O*Pnؽn{Y90|A~y>!? z$^W7nUf7yp׾~={%:?䧕6)0y̸B-/N̝SvJof:O*w _=TY2lvp4~;}ve—myq{t,dtJh|>_?:2?_ϻßtJå[A|pճSW%oCIu5SoS֭z{;}>u[\]k+7ÏveRX_KۿrJYx4WBP(5 T6UlV .vſw_OTJ@֝o&aq3 ,S,ЄQr>| ᾌ'aAc XhDb%U_/jbkۉwn tZdT9T6F) )Fp0u#"[8kOgW0?N_}*ÎM:4M+3eW0{ &<&X7Cy U˳Ta>ºMP녴-GNKO:)9q։k̀|ܰSP#.QE}=9_?´gUwh 2?C=-l;!-k~̣͋d~1/Y3ήi^D~Cz6X<h.q9m zp@4TGbPyEH[UEcGGXw|Υzzȹȇdn"eۢH7fz ~KMpR:q(&%Twßz"??HXjģe1j$*hDx^_)z5DDIWSR!&(ŨP\0*EzI;|zˎFˋzէB`1!eLXόB8UC͉,x!9bO1Zk_PCT/^P/}ZowvNfϑ-5ۑwMN#k}6/~A?LcBX0Гo~qsW \ c0pi y)Kgk9N=O˪ mdMV*|/Y qB =M-9e˖|7LALwR;p m2AMZ>J^J3NAkiiȊHҡیNƜe":Ξ l8}L"!},=dނQ r@N޼ӂlo!ǖBEB@NI^產4@Si@H9Oq^t,/jG{*ӓ5vBqe˜D9HvQa:sxt #bL r+<9ptE,U~ԜBP*x>8Va .c#Vx,قFL:sxt # ;V(ɓJC3 /_D6*ETãw@N]y|MBPJ NC@Zk)幱m)A7|(;=Ol$i~olf\ 2,HAꯣ 6]"fSA~vVS%Z٢|%hN?6՛OhkG#qa`mAC/\Ktv*Sw*|(Foխ5y&z #)Qȱ#=H) @ãy@gU@  mCJBFuzo7P<~69>AK @J###+1Lq Wqg\s2o8T$(xjH kL`!)|v2Yw2|v|mw_rxosQYK(ƌQʣn؀fyQeGVSuc#uWq zI JhBPĠ@%zb;#m\sE,Ioq^a[~ ys,-x]1z+n7qU2jpv/gi#j|_M~^Oo=I忦vv F9cN=aRp}J- r`)8c@QOUY' mLg4ϣe#X0JaZ Vm(ׅ6| _n´GĤD` a&LNgP" gSjLc!4A]1D(8bzho(dP"9cw+y97 D 9a@)((2:Y! Ǐ^.=u\c(ׅ7O8-s >[K:yx7d2AYv34k^?0|3nP@duSCf ATPs˿;fկPk*oj,uߓ$ kh_bR1'xuC Z6O{@Unj}ʳk&|!kpp< nu S%z֔;!"GRd]:vH*c!,vˡ{9#A;e GC@SЈBȇB[ͬYw^0tns<}YV8V[vu .R8&:ZW=GU"9烜 ;|:bRr6C%7qU$<$|(7GNq7B:L]GI9QRq_t"GJH~}wV׎hqV2#η%'>(l-6H@I\"BQ]6]nJ,j{gI4Tq:H|vxǦOǩ1³63m4_y[xre kqrG $ 2Tl2],|(g6lO˹?Vq`  N|*} ?$:rR L-~jidq$q&wT+Q1oNnjmF"o`K)q̅⁴u= ~77hUԟ7g7fOOo&a5IzdXWub2Ieҋm^N+!*K|'$~E39">~lqn5g%{҈pirW>aRHMɃ7iH˧V, ߦz`ME>m ./AwA(8k/F"J}9BE>f}>~ zZ~OiJq9cw[761yXN"NgUdV6g!Ph)P =lO#d)nN8;'%/1BN|"e?0[i(9j &WY|ݎ,K,G XQb=[<ϟHT%B@}Sf z|SN\a沁{pE5;EY:s^+Ǎ͉X|n z|tňy|C=F9bQQeE-CYY'SLc~C*f52lʙhe|jm䬏pAxWwn- 7<t`eBW0)t7' }&{崔G4Ad$,n\Yʾ<EۘOʿxRpFse]qhhr*r1`\ /wB`G'izYWȐ7YnH盢=>Pb[@,IDQhyi{*RC|w=rMw(xs+i&} c1 q׿Vr>4!@ .qJ6mz|F:M%j>b1d?,_R怂쨻NxP1"۰Gv}9֑Hm'3jdKBвz+OߓA }z1܆!eA]qGiKlG6sFNB>D=>PjĭZtYf:d-D)d(6vV'@.ڵݶ=R3UTh)(ZEJ3bZ/'ϣ7[io,.|k:e8!MBĹE]~J'WP}]&TႧzdܫlmXhCVs#oȁc>Ѫ&fz:xXf|X?JJ OӖ )t\|p4@ !O)4"v(QQUNH7ؘƕ6o-@1QJ&sP k TpԤz@Z|z|dtF> (zldz]RW2Gc/d ++.My/O6ળn|9\ CfӁoBc*u(§n[ @< `T/ߦqlpncPpRjR=L "GWJU`H ˃Pz-v'nUVe*NZqv]MeFܚttANt ok.iU:n ,FձdCzyl:xo NOX?^]*βL(L-Q&Av<@A7uM uRWnչ-15dT p dq 3@y2KČmn3 ɑe508t М@y^'\L*2 U=MI[&7X`No44ޡB(rAUO˫k!C'|yѣQȕ5 &XJ b=mnm|GzʌgDV?ػu"]P-+xbsLL!ԫ/ uK]ݗa,e=Zqhށ*qmxyu$/'A>T˯̠A1i#lq47a"7.6's&p=2̄ Ʉ|Y=>TF4?:ϏXq2OSn dG͆6r+E-2I$5FG.0b#ʡta3ͰP[7ĭsEP#I;TpYKӧ^V=w"ȩ9{:0sJQ#h93&p!W%F1Sj>L=>TJOpH M5c5*@;,wnPFM;ACaF^^oEzKte`4@k>ᄜϲHTFb-Ufz+ԍ{ ;x me; m{޶tACr(Vx p}BQDƓY{(@(r)/ ZSelS5ʐ<ʝ6ʱ - ܖ~.B}wyi\rGrVtEA]OQw(7'Xn"*9sSls51"I>a[!lP٪~bIWq"缛rqBȺ~B>u@1%[56Shsb0=A}<"1'|, A[GEi/j+]f4%.`˻&hgsh8q|&眄*9%G"|J8!cf5j tgi^L| V,1JVV H<_ \a+gw$!֛e]o6tAhHm5CdCEkp7IMHT)A ѬUU&jwO& -& O νʠČxyY?]}oν\q@~qWuu} [K]VVDCr^e"LP+V-u0( ()D<EJyN:1S[/(q]Ż@] }(#ȃK r9)(RD7JGonRi[d>>6*Eu824LCEndI[)Mٺu Յ-[VukQ>⌅}u95NVxi?oBp" >bkyAJ~#,mgQ[Bd0 ȃctDya_]_T`908ey],u}OАLJYN7/AU%<)6[E?q=Ke?Ue} ~Г]!,aO [Fgeo(}KC<oB"U5ƠLv7].!nӴ8ozS-~#nN dRiiTH󋸡YnxCYԵ;m?򜫟f/Ej~Gl)GsTю78a:tbX@Hr`N#\6~iM&]Ou Dkn=PrvM"qc,B'zUOl'[49ģwQ06LcI.<'sO(6%cc&Lj/ LP2d_M6P١qTRDcW@7E7AIݪ9*`~x4ݚveɰ F!F$I$NE}7t t:#նL Rez袊nCd dʲ VL=у~vX껻]^^^Eiv'GkA "S:$P}!P'W{ԣ$-I3ug:Kd:ţ-;ϱTigw0 !;(AC!9N2ytYX?;5I0ͭ{㽀7}Go=Q>={2"y\*K9M.s#"tȎ;^6E.UIaBzj+?=3Qūӕ3yV+leMf+,n-ab`kػ+!(4z6A*9&~Z1f}!@ųȧ/V?tbK6?~8a.`U*}UWdBM9aמZI%^^C\{vЅ ycwG{ȭteĔez ХL](5H*]&(HF[iFS@OfurVb"\0C2m nZAO[P1 %y{`:q!3*(a7C&cϫNzAM{VU sF|@,Y ejG^@WM7xkl < +A|߉F89R#oPnq\ ˇ?q#]!$5z䷚wSy~[7ac vv=W1=[t6.7qW8O1n79Rcgn"&ֆV[ލ2pu stT SX*Ҿi0 >>h}GDi:r,tV<&9zCx8*^NhE 3cACsCqeV_1s"?q:m{PsƧ, [ЧSca@ژr| c7)ؗ΁+{&٬R5HkT'/'3\5Umڿ&ף2QN/5nF tC\2Y8bC'EV}(2ur*5fŧ?u;,)eF1WjEx/y+ mF69ӒīT=@?v >v PXgTNV9,lތԹDU%܊Tfx3m£[Sm, cq)TuC/s |R`XN'!nUt&,?t6G#dBDe ӝD;c>ȵX?b H?Äc g?{89f(q*QS9l>_RK:&>8YⰏaもc*7I s*Hdz?Z?n")# E\aѕOXk5K: `BcCGGYo0R,.0 $(R⳧I*?u8'pO3C RĈקCk υuW8 '+pLD5f"9CGٶ1C8>x,:y f]NsoWZ>|`,+}5Xɸ&M4Ro m <u|-..wx4ŏW,lwὸq3w^6ETCV F$^Jo ;nއ dm+#KԶ3Y_[R˒*{l'|Pjb=YUB69*_L{n(}ieqPR>֠**٣5AP3ߎ(!F1VSTS-QTK%5 bA f24PԼ8E7v/* l a6Ì# xt?FGji)uɁ{bhx$#z0u1ACnb^aw*н{]^rWsHrU8R݋k{0=3p(!=8FY9'XxҜf!)5w=B~Zu*bQU{c8+WN5’8IdJ8g8^3^ܝlX183ێ(A7HT)L9.RsEX>$(HťQpO .\;4v;1{~=12h( j)#+!YdDQ)k<VU]U:;IxirDjX {;a!CrOe]_w;1hV^wC͸Ԑ)xLR9гn pBx^5sKyL\E$A4#A23 k)D`0 Y; M1h2k(-Ĭ:ymDHad"P'A>Vr:|}͟CaU&/msSͲoTNPDoIFMMm)c䶙Φ4 <ĸ1R~񧦖lL&i\ kE2bgFZ1ڧecF 2㪠|aPu;tFUXP Z0ͨA54kG}nP&yMf\a\]_=zAׇi0XϯCcϏqau9a$H/#2GѺr b ¶R=ns4} U@V`*pP+-;E wS^\~8󫻻%tq=(/t>k^>,3ceYG(֕ ;jHgwDUE@SnC3?䅎?6yA<~,\>,*lg3z_~Ψ(Pnkʠ(h&@Cez]!'dm2b> /h)5cگC %i3dL6Q //ĤQP#lO4m/J`O$H=yY_|G5v퍴hRRJ>xhxYjV0>>8⁥ԂbTHEdvq'mŹ˙)\/ˍ j{?`R <,;k7 WN/8Dy*Z# r>w\.bV4pqFM2F'cIr-_YuNxpxcO+e'5؅7+c⑘XJtYԌ<;B1윂UK`:!<zk  0Vvh"]ar{hC<:GQTEgfi{s5,V[$"c]MX'+ a_e60di?4O-_V44# Z eD M`$qרUoBo!b^bM4[4Ƹ1W8a3cmr:YNTK(DSކx! 0P߫`q Sޢo<ބ=|q|*q]0 +xx.C4㰯-aV.f> :\fx6IbթtuMcQ;cg'4`i9 0K,iQ7o<m#"jJ-'c+'(E#D.WIJpP(JPU5ILJtux ` 0$Ec9)xM W=()Wp\h5-q't[{pm;lc˔+.1ʅs 6[L|N[̸HDGq3tNߧC JRwXB(pjk51ݤ=NōcP{BۂC٢.FcAjkek`Ma_ލUbF׉pDpFrR N֌qoj^[mCm1K|C F!)f-!]hӚXW&Rz^ݝ1VjGkvF`: 9ލ_OrA#c Γ 9-0 A?EKZ!䉫XWݧ! i`&fYYx}~ x0vbA=ugqTNQ)Tꘪ2me}nr,XRkX:'*6x fSzjhcPF8(WR4x(\( Ղ0*TA:- O1F(K9VP7,?Vb7)ےP<)X€B7%|K*ZMI-tq4#I.vaga{< )s&JCߪ.~pQɂB͸Ԑ)÷^uTty5͘P2u4 @"G]RpWxu$tWY{3JY|}=/V﷍W.71Mces"01.&E3Y)lU^ rcҪ7_^\LX7\l.33n/oa/'xxH—\ȓxާgSx2h;y:xa*'7~pw7e ULAo|Rh\B3^YZŒx$*?FqQ{+[|vWRpQYTuG/W ZKԫ-wڰك/.,}8#r@jdO3.oՑws;ǗqgT,noo/kj1oȺf9~9 g_Riz(]ul3/>f~|Za~;}}Kz}B?\g?['/ x;6דZ,k﷓UVEWbL$I &i|[@Qa,|^Ol,?GX*~[ώK2yMN˞\|fGt?d+UtwNDA?מ+֤A[je, !8&ebl:2_w*hoCg 1#aNA-ϯwΫn 8qGt9x1ɟ~! ܯ/6O7׻!hv, #,Оp5|-KݮJuZO?Mme}vX7|aaھ*#lƞ ~8IjNaڊ|)[UB k%A` էǃj7}f[}X.C=~bQ̧ggFJMuXo7N<(7_|2^1ֹ6ւqTPx)ںޤ1j^}^(܅F5 4J6Ova]5wft| V2Ǐ|S<Me2Z\$TO!rೖqXUI0SA2 l%r|>[i4Z`B‡XH+lSpQ3}9(F0ufCeeI3#*0g4&?)xl h>sgN{INMs 0E ޛxz,ptӊ^o"2x@O xu\.wGxa7ϺTɢXv^ lS^,Of63 y޵6"KKA<$v`idOQHmG#4ඥ:$U"χwqrqnioviPu`g0 p x s4&0o}|KNb<]1&~a8%&q$#LB9a#‹}K' ꕀ"ۓ0z=ksf!*p㉝?qĈvZ!S-|;C$a5f+f|5QcןI鯛-MHoVnw8X=6<8g˸|nL @mLlu<] t{N-zu-}D $;s')1``s.HG JJ=r?< @[ϴ"G KqTRD ;O uQ7^cN)ЫL{NĈv!Ă2 Dr v(X5:` %ӷ,-uͅWo $s~>oۆo{|{ /(!p$3CB1*`3*s- "x d9cP,jg`B :.nd'!,:V|_ČC+KaϟRfU>~K*`h8C { svG& -i/!>X;\tSP{Xw~tK,iŬWIU S߀-M,Z0z\l+v!<XPW!!%S`JY^9)5kJҠ ?sic[.]SGt !˗|% E렎Ot; >ÉV"Q#PYJy7V@Dvu6c6꒲Rs_t}1bcp=x|K{L]`qԅ GoJ]MBH=+i~,`Vb0A q2Jp䵦VDCwmk4hIѠH.:0Rcj.İx$OjwÁE3HiVW5˻󲄆XvUCSх3NAk['R=/!*Z!y=uz45 EAE~4Ia1B HJfh8Ս%=n_nTR\ulws`n Nz[E{KO#%{=Dn)\5GrLSD0yx(`! FаY cImr YrοBYXҞ@\YV䰶*iLvnGŽ.][G{\B|0ugs`}ᤌM'e]wPD*qe9 i2uCVFH4ܾҳ췄U> gKQ*+&^!`4ƈRvuݓff䒺9,oԌ\3*+ )H0EkrZw"WښYE ~DZ9X`Q 49b%RFD(wI۞\_ߑIUĹ+=83 *qI]]ʺыJ4^ ,suT` PlhL2caQ~:FjBM661i)(*쑑#̄  % :}6}Pz8Fd$DFvpr Ř!'Ֆw"bi)NSO]B|ؓBrTQc"TA`HIaTެ6KZ_3JզV  h3ä"!{>WOԮbS_Qe]qbWaI]>l0('"GRbw'x: ]RWOF?5`! <\=g%&kpbIE] un UyP')bb>׋x@g pPIk>_pc^SJĽHӐ eT+=+!nOp̯zIT֘=s'E͜cWE; 1L*"j/OxW|J;+OZTttR;az2z&*̡eMQPYz PYB|q[iAŕm>?b|I]腣-Gh&]RWF/U\%{ק>eqՍOSW&1]"DF сo]`.}EN?V5hA-ZB|Ь@}qR)bAUw\ PHcI@ZXX  `c^Imm U3GEPBb-LAz?Tp:~zXl,ۢۢR+f9>Sh p} `+}Mf&0A~mU7XzmhP$l}gdcYu2[EWv"zt …)Lh%ao)āEHeԕILފFߊFD<ɧQ\/U1b秲;~\o¸8e\! ^އ`<`zcnow[}aPk 3?>#_]{a<|5~WfC&7W)7P>o7x@Ť/3 Xn?;y8L#oѪ#/5~9>x|a\^\.dvHM0oٟ`x7 = W= M(* E{0t1Xҥ9ԫsCwz@:g'ܦܛ6 i:|YX"xpʏLЊ VގF9n1X]g!uv0hhq>}'TaG>=9yg s0Om0׿ '4&=ݜӊ G<2[)Qd{Ijxz*NR[R7PP%ĥIja5+vN%stI]ԫ{}pTS3 ϛ9PpSJs6M0RcEO#+Qi̋N`YQNj_e$`6FQHlxx(ՓMז2aKWfgȥR*F FI 91!=8O?G]dzƭ`D-J tӒq 5kZdTB?Bk*'\%xmBT81oʒAI)EI)!.uW]ьԫ]T%(ky8 ^'8z(VA6\O?obt1հS$!f5 I3T8;՗і+!bW _SXYGZC/du΅\[ MFYxkˏop'υA|8 8\ KP$ xΠDo++#z rI*0g} Fɕ0:z&|BR25$2gd:%"xbvJL`hRxutpl gKX`| L-MpB_mؕ]13=`<\&- >?`tylg#c תY<u~ɛ?awdYAׅII_gb@pwT p/˟w|0h3GUF ryA;-%fO7N.J΢;%2Xrl.fI7 *)t )+"-2VkNqvLENr56,ʂ^p~כ|r$Z5!s&bbIb.*N7 k=I`#*Djv7{}3(`mY/ i9XM6ڋӜ@l<9.\E0=\aA n9iDb:KBJW_/ (h375,ۻl}ӭ/5ݫ!.(lNaYyw󽰀ȓK2Lؘz 1"y@0b #!45܈ȵ@ CsdL)顓#a~MkP߭7ɯ/=mA$ _R5]-9ZFADn"v-W[hdtrzAj;׺ ʗ,J?uSOE?'~KXE_"HJ&;dܞ/{HbUٻSZL ;QM!O"-*Dc,B[ԟnâ]4֡W:Q\dtg㒵V`^mNіӤEZx Yg ,Zxt" Fw#-tc z;1 .#A49Sn>Qs4] yBu]罆ATu ݂E/e z1 BML܎vJٸ] .z54YqI[c/ ^1(Kr!(a7QH"৓;n:G&7Oo ⃡Y_P ݬN[Zs(E(jZ'|y]Y hqkR%$~њ5a ~x٠\Hv bh詨rZA+onE/3G׺dU:RpᦼRT,+bL>k~ŠX^`S1WP6Wo ׼5%%Y/sYkj~i!jXkW\Yt,?~~Y)qŦ駩oA9Hj[H{O(C\^Qh6<.bq13JIIzۊHUJ4f\R# B=&qpԦ@_#s4}ܱ2Z'u*$k5-쬹YΗj!NruȐ@E )#y/!NjɋM65_ C/V *Swxjn(GkҸڰchV!dAG,qTRDy3+nr~7Yڔk^f|RUѱuq}jn˻Vl)M(f|qǗju45hLFًsM72+"tYQ(44J3\RTcG:Od/vٹrgz"lVb}@${Ds$B` PRV% =#< |{p`mJXd_bsJpu AOZ\Q({x|5SwDr 8pZa])VTX:pQ\=.N3./-OVgW<{q2WL3ڿ| Ӳl}͗pRq3I9bWiENÙ~ ;yIJlDP  #C Kã28t}#&ΥdLp+v %ʄ4+{xTAlƣAac#sĢB\P$(Z'Xxe8mi\58,+Dp'U'vBqa$E:qP̅H=hɣ28tp*)thKG& Wh|0o1 "X%SD@/}3<&j[6&öI2.7e kEn Eea jy[>Ϸ7{xT7qںlsDI&Qݑ#79FZ/󬌟꾆|n` {xTGzU kPGZdGzk [Kme/f- E I"C[+AȰc?ْ֗r|.t w_e:yӢGep:+PjH$C!s֎GZX 9s UN=PZ J_XT=gɑuд~f)~_MR8m`0?Ƴ`9L/0P]@OibSA2(BtI8t-Pm goIJ%F#HE"e4Nr0| ?L( I!j\z2inXpt f *E#A8rq֙Z8H8#1;dp,;pFxnZ%qWFn=wL;>*HSޡhၖdWQ61IQA~qGn<4ݟۢ^(Ѥx$7у5\n"ыѠ}*c#٫茢*uYRPUmE8WUw8[WJV-ҕ`jPEP\&}-:&J[g:wII1o"lRTH5jEMYH]9☃G(ϊVG;E&CIZpY}L£:(ݞvmwc J%|<%9 f:$$""B^QJ?6$[θh%?/.iF8z-?8g< {80Mf8MG _!iVǃ'oޤm#])Ϳ2?1ΡD#{:V32quVo_>>m&_5kcD0,>EmS0_M쭐f0o`GC&{ |.O}S0) m\8~6G4_ 7kRLy-}r|mة([gM,ų,^0\~m^B6y+"d6\iwϽyl*E`0dM,rO|æ@_-)s_2O0hJx;'C;V<|I ov0{cW{H~~Q?)7ޮ|ɊnH}nO߿{Gz(l@f7O3!xAG]2hmKoS29D=[Ut<*I{nۂgZ#h% 4 `='S!t.;aqanm{iJ_k9%lۤF @ ,ɲtbgcjKQ}pYqV? pF ,y[ky3Ec =W>=<*Ile|0%= *#kd9e.&d>fPn !ag@ i81RЀ󨋏0+_mΠ(0Xdun"V!Mu@ KVIK!X_.rb>PIQG.aa55Dঈ4NѺł(s=:nqdїqY_]@N3,[rV7kfΎ|58̙ЯқR:7;?1O֟'ɲdU.[ůaxMeHc(8E{xT8gY{7+wp fKEfdKL~nMqLŮwy29 ;(vEӰ+XaŔob*>ƶ'7ExRǸJ0|h|?pDǨI0g1.y/QDCAj0:FNBbj ғҒV3Z ms,< &pcCΗJW޺4Q"xn!{F@|>6hqzoZ6@?ArTzh JMQ}S9 @ ?c0v+0ۧI[gfǛgu:Yɨ ~ `c|BML0ˍh!t]p>@xz!I=Ip!5x?o D1.u( .pU- &%s(|?BxCbze`Bo Iw Xz}˖ʘ#':Lݴ3#CAh QDnF"`GqqQ%NՌ׫Ix^zAQ;w4^&6h@ߺI>"k7ʺ;c``a G%X20uXYVy?Đ"#8ƙj+ P`"-2/Qb˃zFat=f= |:DYc:Ɩ8Ҩ'}RtNȿMZ8B 4ؗw{#Fu9[䤘=xlC>`[ 5;E:(8Q$,t8磡H:F 頧M7jb@Mv`u/r\~(AH-&J]cA&N>5eG)$R-1R'7mϥ" uôYf Gq&:hs \hĔi,I1QzkhP'"8z*NfĶCaiw5 $ K<&V~Fm-5&RxzՎZO?dlM}(s,YQpќw)-NM.>Y}\~tAB9皿oҟIRgBL!r*]+*:?cuFTxz"~Mlb =w3سt+כ>uFGh-$U@ds@사9?$HMq) 5u($Fް??~_x5,xxȾ WTrHJ W4# ,?x P+㨦[}+ 8)B(n(^Zc~=Y6j{ǫ@J~n/~ wB(8w{?͟^)} 󴁘e$&>G\#ZZ|xojtӴZ E<_7p=o`l`ap` +VNe7X2o V|>ӵgΈ%CT[wEɘP2 RL#?,(HS;`kvU@a7 4 Dhvnrތ}::1і`TpM|_>*ϧ* _ZN7b qĜ ԗI.PKǚb%HBj]*޸W5GtxDŽg ޵u{x.3&#q+4ZqJi=ZzŬZ#SJeiD)H| }:g_,~vu̻zL&"Ympۛf#$f$YIJZu=oltM.Z^)4PFX:)9q E$QLQP0r˝Q`)${%N/eRX2EbWdaH x#(q67Όֈ&y) y_:Lg7>60x~"8~jRp=Jp+wpYdUVŒl17B,?G_ dzM`=st8 *p)KKmh$  ylVe_ƷS [2{WQz9SiO[@BghA ܒHm^=tġk xjuKP/r& !]ˁqN*zjtt޷`Z^S3-%0--xSF% \gZaNJiA7EA q(ZoFz9AFPG,ӶԊi?ƅ@ x [SSLS,%P]4`{S7Ke}`D +'"+TZhl-Vԭ8c>HF w}&g\Dp^. BAh73fv3u-3̔(Va#Gt 9+ 2 `>%qڭD3n[@-㩃H&˱4i6i6gթC 5e$- ğ Nʰ<-5(Dz,pP+h xiHQ}IJK8M$W"4Sz9ת tA.HiPTSKM*,6N#eFzJb'T "z%h>m ~U B eQzD0BŒ*0{`㰴7^<Q#{P;h hi5DUrr$B6Y|򻳐E~qZd|Z aX?lÛ J Ɩg^zJ\l+$ |qf>x}S ^o:|e) S~tc6Owlg?pvO۳pt}C'?蹆Oڊ?[`O`a8clϷHAk[uZ"ZmtC4' fyhN N?!Z%i|<~va;CT;GcP6 쾮?DaL3#0)q sVϸYȵsk.\ A$*y72*7mbMDby!Fa^ mT"@KH!-\FL e@G\rqB<ܧKQPKL āBE3sl݁݃?oeˤխw6L w *LsЙuea +Sn8 !3F!en8jF&KΥ w,+p &pq50qgR`B`Vɨ-WXBa ȫ`ࣚp[=2²S_,:;Zq )^D$pYb+!3r0qNIrZn`*Ln/% ͬrýK-ɨ>0 O>a,::WAǴRbW>ʃnS[܂#̟u{P(SȯjOE5_|5_pƐP{TP/66~56o\QpN8*(U:&8-+QVzj%?]))b1K f4@Ur"0TЎjñ(ݯmVB WmEk)h6G]׽c7qD \٧$- )qM>)`R[3A4RaOl84d ڧ(4aXۜycRZbK +Ђy LX%&Jhi$,*K6K: 1ךPLsȨ F[❌3Ųv=ЂyUsfpA b#Vz/3mhvB -&i Q%W WR,4`> -&XCD"b#}ڜl=ЄyG[ya )6&b&@nx!Z09Waz2,HRNr"NƸ7=6?Bnͼ6d=qE!TρCL8 -7B̳d Wsᮂ!9S۬2 -BGNROi\Ҋ$ s.s(4aDd1oL-))8+yQz@ 3/NRANgI=\ 탷C(4`^Q\Z(RVЦ Mw5ȩ3aZL*C349/^!Z0Ϲq͙BD-\}Sh`T%~!A!-\ \>)4a*ٚy"5%)x  XUem%XЂyie뙇)D3p)gV3$u)`Ђy]P*&ZW଩'ޒd { -ƶV{eA .I.BiꨭQ$@}-RKa~Ś}G ׵W@eZ1m# \8!Z0okT($!}$Ĭך 'LЂyE ]T'%1X\8ɜVgPhܶ^Ӏa [O'F/0oMkxᬐ@҂@P%(8)!Z0τ9WHס1*7K!dXқBgnD=B@a%Ǽ$E~#}kn'_SJGк56(y]ɉX$E0l%LYB7GELL{PugB1O7&k4x>gK3y0{ -7̶FEVUfˡÆBaƞ!Z0oEE?Kc~;%[p\Z UH䅇gBJ:'\b3V ,ʧB̊֯x\JhHL[%O5PYJMIZ0ϥk yʸ-@W$cm^k  -ȳ[3GHUqV(-BuaI,Fw}Ђyݔ !@p&DOd`hb}; Ak` ã||/#&R+o*t$Q!M{}IP2]0A8K`@uvˌ#5v-$TXk*ˠbQ¤h2)gR6&T p3I)UCjVcݠX+65h9B xؤz= 0& r&()2*S#gge/mwH%g{=:Dŕ4M֪Ă3Vp W6&\yӧɐ~ZYuMݥMGwa@!_)Reeg9 F? syԍlZf\< tT$G_ tX\߼_)cHQ^y%̷!ϣ2}>/.`ոvFg2bvvU4GC/OVj o9/+% =(; OUA>xdw'/e2)[˔yVDFr+璦 /XNHlI` еtE{S .&7v: *]OWMo{1F@~(p‚OF|e1SngO5-V:Ѣ0"k\`mi<wa<3zV#])9Fg;a6ǶhmD3pQIWO@8 Gas0TG)R~Cq|Iss;X^+zvֳϾE<ոcz:o{ 7%Ŗ۫w8m 1c6gt6qCM1e_oj<شvSZ#~v&s@>Rq"7gUXhհl?P6u q/|SfTqB&Ulg|]?ZzaHa4+vS@~f7]g8,^k2X%gA^;LZ`ja?5rq]>2wX0Oty'>YzĆu2l7 C'˷ZJ6Cqc(ڶqզW7'/2'/ܝ%eݽ(,{H{5$:1Uf-x!#Zr{Kam:ytW񇕉펙yH:@OIhuWz*C aƻX͗jb'N[ ZR:7ʫBQ3F ' Fv܎RClg`1mgmVYONjoys v˽h8?{9|@:nJ{Ld$>yS1< 487E?ջ 蠥q!j{jTHuZjҜѷiV_ӧӃONVWP-|L&gWf+:|?؞ -I2kԋ?W 1Uyw[oWa+ĬVygb*lvJޅ$T)Dj^;iNZ6;;?٪Z4xL[~T,; t˲x9z6OmZe٠ޒػsuL_/r%ۆSw:;ƴs,٩:jb#%{f *ja%RBwEs=pK@SRDIߺ;u]ՌOe &Ȭ$*%A )ٻ6rdW}yFKb43,RҎxewX,I*lE*,}DFe)WVIM_tRea0/ MY ˸$?.kEkE!JC\)Y #t^SW "EB:5ɕo A'K%(M fL]RsQ| 83ulU@*_R> !{~Dz4@6ՠ86UwRGZ՜,\c~9 /ɕ#:r7,堊ȫU-=0ί !w`ɞ-V+6$pZ{V3q,[W'?T˯=JFJGZm23L#9u#;O7~Z-Рr撌mTk-ipلgVěvv^hwOQQw/ۧ:[ E5$)W|VYU%@K?f}&5oeyEY!ݷe!Ϲ?Fƕ-o ۓ~:?zzNvEjL遮?]0Wx_%ts Mܸ4}t)`P;-mIlp 瀑qIM+B0r5~]B68M4!&v2.CG/ Cf *HU$Nv5BFHB4Jْ })8_ܳq! ºi묭Td40Ϭ1&d?A7H9}MeOM)hNU݇oL,+numQW0G7QQ[nDy(dW#$91 "aFު\ӡTy<&%դϓ LPc7iR<̯)wme &ϧڕm/;χaٿ\^`|8[RR2sC7̦=0cp[4F392Y\_;W]7q=xPm935ېd6)& H*)E2! 堅 BRBZVIV%혒vvhz­Yڪ* dIU?T"~~ s<6m[g朲#pE|sOYHC! ^pG\OX{s?՞ =yzcn$>#g(؛scdb筘waz⎫}[m[UQA?a@ڤŻ6Nr|ӏo/7F\;K@N'ԺS)%n܆N%?,FҒ=&shU^6׉跪v`@ɍySI'&.ϵ%J8HS ǵ ֓K=FA]`Z|ɧ`>1Y{ 2:e1>=N"qf6Lj\eo]I0KhMtV!t :R&!HVFZGtJ TJмSde2#sV%* \L28lF ֜ˠCʁ\uKp'κ9n=98֟]|Sݜbc?L̒)j7s6˜ыkTIr!0D(X\dXʕI P [v ǫp R#Vy^7L]4YYYbGkN!JC\) #t^S)X %q&|;Tl:>@[)̙(L(S8΃QȟF8IN%dN%ahF:b$nFJ[:.8d#fȼB#6T: a3P8lɖXtzcOIFs>ԸSM`mQ! &pKU.L>GI NH,0C5쪢΃'<5_;󸱁&2fAȌLHSf3۝Mʓ /ܫ>h-3VmնJQU$NJq=a0T(9j˸iB4.hD㝎soL ,X'rjg "ېd6dO19XHQn4R赦&f^簈ϓ6K_cr' f=쓭VU-sT\Zjݽjuy@KjrPet1{OtMl"P+d1[ϣK N1 .%,*]k[sF"ÁuO"@{ ~NBu0"a]TxoP:( 6؞9V2 ڔ;jeoH²T! `1e eCXưgF@e^*3M: F Vi:4ɯ@ ,QH"MuJ9,˿_|_UJ_YΨxuoM^0n> dI/?T,方Un~&nn`pt[)>W 枲'*C 9A(FƟ2ԣ xo'gYyzʬ4m|ܥH SiO:(I9 BL),pQ}.E|.'O7?dyR;inԸm9L `9: )* MpVuJ4'f&lmQX.ۍD;,F=&s~heb[(v+nL)ͪ:=SF8/iB`4%2i XE1x9*Pۨv\C 噭uhW^h{pzMHPFzN.`v~e:ȷUWwR 衂>{yG9s|T `?SS /%*:nX РK &!CtR+e"8˒6BL(93 YD2%wm$K~ڕ&~ii3#F:b4njFƘlkCm(|1Z* vJ܅ }yc} ^Jn\.&BA=1W[} GBcʊRqrQ[$bVWNMbLO{I%b='`!%ZS\2`'*p[ɚ8kքx]%8߼3&HyyGT:J3Y8q-5ٟyٔP} [nף~ǯD_[cEG'BG 30 3&b qobG/%:>"4߳Q|G+`Ap E0@?N`fY3|Oaw{#\ϲԛS?\~Z SA-4nϳOKf!ާ?`_kyß8}Mf ?zP ׮ť{ r4~_L5Z#>_>01b\ +r^MS^b*:>pt6(K6%U6]B _rO4G"&Pŝ'^)>|q>rrZ†h.#B,=6l8Od=K@D&k+$R1sdA)Q.s$W6{~~?:F^?:Jӛzc gxԫsxu|̓1+y^ͨ=mΥuͯ|6_4Upϵ*̒-zS,`lo9^^dUA3);S6ov`|p!MH9oQ%iC^vef̪U1O+uxbŪJqdx:J_:{p;ٕnԛ\ji^ vrT{35B`HG0,0??}Fbn4\C`pO"-y@D+X&d#$Q-FaI2I%-3!YĚHu0-+znn~,42V^ O#n[~VranM<țGa6/>|m_Ow^~<ȾK^HвBR~^Y%Nh06SG˩) J[*DVqM-TS,-`*%Qjhu\cQ/s KPbf`V<"Ԕ+]J L5fFenGK 0(0qhx,rSC^ibʃu`7g`۰-4}[mwО}V^okN%}DH(,8& T*`g%gp#\kajp`zDE]1"_PZ]gDTse%'ΛY(Wl 0iKv9M*_$hcX7h1b {ʪBT*@2V2 c! JxCDrY#!%62nu4[mp.AS4(^2 J<5tkZuz[ Ԏ.VjM]lqEݱݴc+c Q>D#]؋$w="Ƀ5P@ˀUZe(Ma]* 88Y[e qb-mNM..EP20)"<*% P0͌Na,`ApepPB[ 3ÇR%B̾v^hks17#_%a_~T l1"y@p eKdC (ijkekRؙyJMɤ@\8M-6H!Ud$$DKehBJ2P9A؁Wyupz5_ 1O19 gP>Df[쐳"৓;n*伎~BtmrK% u] oy;oD>{ft%wnq]~z'zp%[Xo¯6869X%݃)sFv`@>AAy#;dm?^"ǩt_hc"Q'o3Hh!Ka4TKظsco.g?t0XyRPPTSߍ!A;B~dč^DIC+CHPd|I3][z#w72(ډڗq^j:U4`%W}.W۷_L19tn<2%-,Z~|niCز=r+d3Re3v3/&U^G/=FQ(J!$p*9LOkB2N[$YB TiXl:-fr[XL2k Ea[(-<8w;fo6ETsa'Oʮ_`2]HFa94be4[4&I5A%İ6rDŽ%b͖<BÒB$ p`2y(SK!T@ B{ھ:VuthqFDF1IB)+ keD]\1͹:q P5jZմCjEkbKYgDm.8iBkvȊ0'JQj\<ގ"Rk)99޿+-*$#h% 4J s.ZҞLT vٖBA=no qY=?M SVD*dֈs,"'#B}ڽllcBxL,9 !-њQ;4PزtY˴&c6DL,IE)1Ab s F8BiWtvo`}mM"Tk^R~BxVXщБc7h8Od0db[,qhY! )R Z?Dq Jof^jXM\pIR)5k _y3m5R/c̐kU%[VY޲*Ȫg>vtm8P4:Ήr^W=\~KTW\U1bN1W"몡zĊUJ1Tu,tL"*v+67ҰyOp^G$ y6.t{!ܡZsr+[ҽckaB& `E'GhĶaa& 7B>b%9~U>Sjij* H#欔LLZ iGœoPiYI@'}}@̹EuI=q88)79&K/umtT87z6kQzJ! {97QQ/*C^,>A4w?l8/?|mեr?^Tk1=X!bx'ZD-}DBH.xw`cq2Jp䵦VDC5:('fGỞ\.D>Vz%V H%Jrّ;,]iT'TF0aN@/qC} ^'2 #Ƙ݅$26$cjBkwz)HO]<W(]I:Į5_ZviiTőVcx$LUr40ΐu ɢuG'Rݟ]ksڨBy,y2L kO!XGDRmS<#[lJ <_q}^x]1*j y&2 7QXDl(>ƣXu}fy&ur MF>\Sor;ZPr{i)sWp& F5xÜ5{RI& D< ]!!Pt/ɕ#)!!aDa5$I1hH=I)NQSy&gC&. gv0\.jmYXkNkwl UT ^iQx"v>.$J{T6zKNƖiJH85I*ʃ ༆(njqVQ V [s|IZ'iRZ [ L*jk |?'}nr0X'Y钕$bxҚx.rF`BqMe)1uD%[:؈̀ Q$  @ Dh ZzZ1*# 0&"=OA{ 5 &N5&t &eF2 ,Z#B{ņ2g謯f |D>\D2y^[dXιT_?;0[QCBDV_"-ye>OIc]"7f^W(4^3TLv0tbWEA_JG)zm" REe<^v[K*g 8RD +|5gue, 5%QFvۛ?>şQ&2|/~tS7üY0j*^7XIVBWThHhj-w~zu׽G,ǏCm%ܢ)Y9 cN^īц,uj # owSƙq2N!RBO_>(&4D pF$[F/z?N@vh_n<Ƕviɳ ˶b)ök`3jR1{WSoPo!a/޶IʯFrN Ro$̝O:4xvTPDFWIW4i`m[0&h1$yʪR{d $b&nۑua˭x3Ը3ucw]~60<(o{zt>v=GȌդ-l/uJ vy|x{ ZZn`駛!܆[DXϦ-1;FҬk0-9綣]l[2t[Gɥ#nHA9.,ʧc#u&RW-g0E#B8?yTʇ$J *,3ޭJ Ṳ#Y+5xw\@u0>شapϕ'?>6bO -IrN\^+B5!)ReՉ'c"Nxo=Tf6u &ce~het]¿7@7(A?N#"?jTZ8NyU )B"T'r™*1og1gU!c}2S:K%(at\`54rg2D P)1Z:<2r3̽>Mz⻾N~^dq0R!ۼowޑ\=. QJځJ|%8p""WUj]Qm$u.^}ܽlbB@tx{—l5wt$HH1_ 9C/4%_\J>{csWUVKX=xX>uݺzchU$Oγ#6+;G^g '7e|:P9my3q!>kK̐*0sNnNRn}흀, ~rʟG7W5DuH9~jzN2WHf'`O0"NB'N&aZmW MA=ʤvoNѷd!k[Ben=?h:-wJZni t`,8\&4n ' jn ChT%ղ 07[m3RS3VYCV`6-faNxsVh& P6}%`δGNYWiXAN:kWw^Z/jԝ8Nx m}ryxLۛ]٥:jbB\"+FƼsF(=D<u(9'ݩE/EEH/6:]RSbK1FbVcp[ݎ}%wI✻ o/+cCR)d5]Ck0hHǺ^"'7JyuJ)`m3QLTC%Bf+gR. nLRK2?(m^rQp׾fZۊG{w`xiZ]-J+8M>kez"6bKnCTPȐrUIPy0Zպm44BDC5yjn[;!%g&کvMX|t& Bj\ȷu~mIz`0~{o]jB{M 9=u=8ǴX85>wv|r1LNEtS7 .vuB}qݍ*_#>֦u21Uu߄Mݷ9 rKv oVӲni16/=WùR/p-bqjW{'~BgHif|tQRUP22>Iv%LVQtú{Xf=Nk7J*^A ;.DWKJQA D7[땛Sqn}-G9n1nlr8ݵ_/:eeP|םH^!R.VPH$g[VXty<jnv1su%m>-9} 5tV/Re[>y!ߓ逐`Ǐ?ՉZ,[Lx 5t]Z{t5- LUhWbϞA%r6ǩՑ3 [Lͣdoj9Kxܸ 6 -2[g_*:/ [vӵzy9]:O %)jWM6׆.l4~_XҜaWWYZISHKFOEcKeC!j aGNgP#JmJA`$t$MjG Tx㩎F&--Xi#1H _.66^~~|pwi/w/:0uމ#8W?ZAp3R_~(qUBX9L+TX*g%n{sc;7TXfHz.7)C%WKBy1ZUL=FGΣ)=ڻFI娭mu|WQh\4nJRA3,2"n`k \"0 @aU!*DJC [ !0,[=E* $,OxIX\YhR /9F!yb Kpl~3|'l !7kb:Wt"K 9iڴ$ T$I,%REi:+} t`oC5LJK/: 8VdKq@ 1˜*0)yH9 (#&,)bp/T"ı)H+X%Os)h})GTE;n49@gGaYqan"[?D^ 1 b*:& &.5̚&.*:Riɩ.*.jj&H<ҒCzHZ]d5EMvQYkWAPk<_VZj<_|5BrjFeͧVg~|j5ZͧV|j5ZͧV|jճ|gɌдSdͧV|j5ZͧV|j5ZͧV"{*f_#pYPqxh "_TI7R_~ï%ܘN`vI|pR{dHx$ F-TB0cPb[*"XR$vLq`.x#DB0*K8Yìi#ƨ,u^0|ݱqñ󩛯Z7f2[^!̾Ć>30잮;+|xדy'Zv&r'i|?cknlBν^;$1^o]lݎ?nww9nt4tg[6زnwM7=?4y8?畖!_|\ȧd3xv]O1_[lӬ9k\zx^_}s3SBN5imQo}M{LDGǨm$hD1V'gjYymr)^x.8VjP"X%kJ ᙲ=Y+)NYQOLAIƸ}b ]d*6ʁ1xF`}ANNȍ]5yST"ݸFy|4>7nѫ2\GKRsF\^)+ɉKCReĒ1 o F:Rﷳj=VUP|[goFxu2FfgK6|4Ȇjnku! :z5x5\^:Whʼn!<NM`\(+ܘI"/*dake| 6KBHiwH;̐3o8-ر=|uSwrAkG }(îlXX_;jnF4|. ?0էH6.$J{Tl6-*cU6 i $ L XP,H.8yQbQ{nYS[%dJZ*iW-a'_+p+f+=SV܍Wnh~t2¥(ns8/nFA$>1 8MXjR"BH j Q$ q&8yњ#ҊQ\0p Y R*Ja[Fe5Smﭺ c3ťrW;0Y o&BDVDDOInkt;?{ uTVn..Ǿ};E%Pѭv8>KT|WCp=VFa48ɗv%tU\y%XI| ?&Hu24J^J;#H)G_gz@g16?fWÛPk}Qz8jveFZ'ck;X~Ѯ&àH|;X,*~f~;:~O vj8%v#yJw_Q%|NI9}P}Mc($W&1rZpRQMSq+kd]+r"E#><58LG'ԩZe*-[.!ǎ1(0r-3Y`{򢷴PEclrJ*:&w=-TY*f4 vcout~c`lf%77ɵ`z%n"m̲o޼Yx(0+M)7';zLEGU@6?HF^0y@En{o rrRM?EHDe\JɬYsó#WYPY(6L|놩^%N6zI- n?ʄYږqYFv/rU˽$-^5Jn8]EM6`hnqFggH;zĽFhzm e\*x`Z-`4V jgemB-$z7'} }KXƓi׶>{? _̏o 综%Cf~߬/Yz!.XeOnV+`+{&'y.o)+"rUc D4.j8b x\Źkn'&1&'==K`=DiAkKFqRiJ$Y-Z5'm,$Θ FOZ#uڕ.Wb7+j\w맯=gy+7<ݞv_,C^xԕ!:k!z\!ƱH;"]$ӟy2\إ 34wl GR{nOW;ۄS1G'ӯ 4A#ƣȕԜq~㨕pZj{5QW6o|ޥɷBmYUF=N-[cAg5IsgユPϮ:oI[,K+^>Eq29"m~|&:++I-R5J}M6|.h(՛ۇ0Є<ēƛKԎU$32{l=8.$}z{c=o|f pֿXK6#\HUs rnk[T0W܃hYqR>;Ū>a-_* (H=1 #/qP!ƪ̮dw;rQSYu 1 ^$2n ?wÿq!:۠v/34Ẵ|C$[IO$Q'n3V0f``CF IWI߯s1U##88`=c *u.Piib ~\;} 2$'bֶRs=3,Ie#r(D; &ƀ\8@q,YìN*Ge/8:}ed?7|RJB(* @4$`HH2eY!۵\O%!HŒ˘(܌3 / s":sO6;kTAp4 3sg}Cϐ!(dhf%\c6DR f+!7рA\¥B CP R% )iIr%qRxO)2:=NPC Ҿӫ8YׇLHPY\PKT;ǍJS'Jyl7{I,U.qI6&%#MpIbF$%*dYJk% a߬ FԇހH8hw,ȸoEGѷ8ה3 <}t-UF$HQIQH$*jJ %G1;$X;D!RaPB Y2J,H`Eg3IPV9$"#QP%$ Aɵt)\D .8Eg$&u$6I%8Hh jȁ^:|FyV,I,Wm AGb@E"b JIX T#`2ᣄN؏8F$-w1DyȫB:TiT28S83һ>bQ[ZT X"<+#9Xe(9k w1 Vŀb6Y}/*EUIb-U D`EdW 58)hlңq6hC q1%-qX`fe=& BeAh!z57w਌#"gz9,WzU/imgA]&YO$*&)R[EmL"#o{NF|)0yjclHYhGdBE&O}UTUR: xDw.U[c t#fnyR4&lJ88)ԙVNec\"nl[o ǭ`I p2ͬ APUv 65cP(qm$_=8/ `ը=ERX6[O A!NU!B]Rl-sKFHGU]% Iw 425IX\U+D44(5J1 Pe01!j2a+6?Vcj{Q3&"h°tj" Ҽ=[n#e%O͚6od9TX] ("F_{ ]{ #@ zum^_bToBl ^*n FYuDhaʠv4BOQhv)0ִcFX8 s^ϐB.ֺ6Q;kFV-5* σt-2Б98kإn,@茚$ frP%F|dC!RPH)E+Y#nT\pYT2N$Jjƚv7t9Xޞڳ4M`QIRh›W kTrV6YAދ.:XHKPF039ڀ$> ~NV%TFXҹdJÍҘ[?y޶[i1ʇԮ]LfӲ͹9KDP]\!tBÑ$4z`2o6:7 ]qsF]Un I<5Qjh Ƹ)/z~9|7G _  xS!9njj?-s m娃U.1eBAv GRDUZ# jVHvoܪ̰n$!>?z_V"bʅs >Hbv%ԇ&DymeNf(h(cx4RXjf'UUѹ?>B@2C{Mڰ(RV'#<4hci #725 AYJ>9|e<(1SLGjJҸh ҵ5Ώ \=_'7FPٗ9%z!56R zx QX;dС 5!WX?)Q#2pGcpcvNQc6DUKoMC,cdiVBMm2#!2 .X,9vQ@0Z$Tiھ5ͮHDXd3l,^m!:^gVB[I+Bi%;}BI󨣈TÄVAzarAE(0a Lw+$H w$S܊D98;|H V<ɓ@h%k$8CvD D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D Ւ@ygDpφj湐@N?ua&k$0$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D}$9@hχBs|6$~jHB3A$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@_/ 9@hχ0-ِ@hx$Z,@_# da8@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D}=$'KɥVri[jڮ_^_vׯ;ugiv;"+~-НԚ(eS~_]/dq^򋳫Y?P" hj}j@xj M/O T] 'Btx2զ`Ŷ&0s(AsLD'},mD .k] n ̏O_fm'H6pu|{j.yH4t.~ ߽x}O+MG~7MwaI>Kyl:Vk?m@m:q=ԟ')$-^m>K%T:ݦF$z+&,Qޖ:mjb(S[DIm-5%1,k/oobQ~"},4rp0Gtv~{;Bn].koKKg'~9:^*4_&◡U퓅Gr[q(~d{us<:E;khߡ .'S|?ԠBAiݔ߼8x7:/;E"Z&6Cz^gKx'U˳F'y׷)[7?OLg炙;m*5~qVC֟-V#jJ/},ײ`b# ~7x{Ylۓzo%AY-vhmrvJHisf"Q5bWgz_Tw寻_٩v=a?;7jmZ2%ܦO j'p/|_ie$e}p#D:ةMC]UBI(eIG}:` mʪT*xؑ!/.Lt[RZ3^Y?RX,AWDuR!u*+مxgd ѪuStY,[HLA2Xwh=vH=V[&v<|͐gj;$UtM*# y&D*$H2rpVHY!>S덱{PD*/PS;ᆘb7Z&ǧp|IQ-/6iz.vLm˕[Q>*B*]Ziy ZgFABdZ_;[UA/`VuWĐ^}cj'yXTE>|:Tn-a;ޅ\sWml JRI_]"0=bۡGvR ܅_-Ib*U@G5TE2 Zlj10] u9O3Bg" HS %2sn疻}ulތ>iJ;aK'r`h$?eBVb%p+r:Y 2k U1h9-?{4!5R_D~u5[7/a.Su{hu+L>u,u $*5bm6H bߑ%49dzsr\9FgbNնP cd^D+.X--S2Զkq|8/}}ͷٴ>=>ȸqN7_ç11xji2M?>Wunɛ; {o&ݔW??} ^V9f\J]"t3aS:sv ޝ,Ͼ) ^'8nnŦe `Iڐdi/rT{ʭҤ-}'_>n[7-L.)W9H$$Ǵ҉dkgp-wUV̚g ޵zBT7R!'%LɍT}KqfL2}ofKM-z]@y.|W*y d] QtI,*Os\}(Wcs58LU&ajTFcʶn2lE *_7SRHn i^x@d^b3٭5>GA>$}ZllsJv`K6V"IGkx\|J,Wޚ6YmεP/sCVAWQ ֵZ2uֹ({*uTg5x`9{ 7K̳w3j^}۫W eJ\-VN`x;6 xjV$dX ڰ77{gpwYnegOfg}yh0KM-L}@ n{%?  n6mF)GVkk[v.[ |Ƭnwg6n'6nߌՉiː>tj1ԩt_kJm .odٶ6=[ЉYr֖{Ǔ=]iȞG)ei1ڇG~xx47s#7i6?Ai8|33kTx6LʗuzD.nc-TZk#Z^~qq86Z8i6eK&aeb&yY^rs!$$g+1MbL|<=\J9fcKM﫮5%ڟ)ސC0GZ#g5 y KC*PhnmKZeAE=+Rz2z mO{H/}w|ց(v"n?!0X]NJ' AGL6] ᵢЁh彳wmje7?w湧^>XEj7NiN w<C8ؿ I8`Q12e[^Ց\k jJ?%KحxmYk]ɿ1s{}m׶ŨaKK Phe: ɽ"CT(1 Fx'|#m5wNaDa<oOap8IG8lkxYԌòdI^gp6tCE{腆&)}qpt4UGiIe c \@J""w\`6ƒ*TJ|ȋֱ(@'ɓ;R1Z'\U5g?`#V<Ʌ@gGܘ̐S.8SM/C@LǵS~<;b%d_ ˇ'mRfs=,H``@?;T)=GCVb91 ~BbN'ÅB & rŶh+gnNukUru LvoEZ]OOn{H9Vm,˂E a6)e1 ę8&cEDrѧ sN*$/t.oD!+.{ c%JZ)u2*QCc n֜g;~&e |9w͊͜b(w,,'355Q004 ,x IŠAʞ=lE-;B+BKb|Dk\K՗OetRuGDṰ4i\xhJL-BL\(šu.ʮDUu Ƅս55\_'tz&8WXC~.6`IH0ȼB]S-8&2:'R]L;3TfGgVb6L?0.i:`Pd{Bpoh4&446琳E:>g"BEL!F/, I-% J6k΁tX)ኯ:>\<<֍ɇgEl3mA+*T|.پ"ϧ=‡z03v^FlڸƜֱP~n+5R1vŁ*FЖWջoO95oo4mwu7w~xͫw^rx8; Y|ύn/\KOyh.q] ͎kN01i䖙[g GRm<>=W 枋Ed ŢOk̽ˣWMxi4JzI@` +6T%lZ'eC"&{.M\'>H*XBs^7m^.xj6֧0M.ˁ @9՜L yz+\QP xD;̟fJl._rџ8\A+U?6oo:z, _#P5`5hu )GCM|g5?jX=5'c5!g%s(\AFfc,`SpSZi:6# e m~J^yBH;id>[c{L}wu!ϤyDo/ y=2eqh&d9zn9rZx {,IBWFx$p%I Ί$`̶0>>Tʟ2~$9^ЗUT^YWrq/EAm|U -J-zBC_ԩ2& Z/G!YNԡb=] V(t6\dv{yɽQ.Ng-'A&2D+Vܥ,Ysh}29BCjN2NVD&ɪ]-gN-2R.*㋯KW޿lcx%- 1+Y>KLY( }$hȝ4 8"FzrriG!+hQA j5gg܍R B W5&Ou!}ˮ4/g0>Ǔ7w1A g.G+{[RdI :vGʩ \(nS DhG[qjZHj~GC\"ⵛzmk;݉S \iD}2H9YÏƾ #BL/h!ĦhcazgMP褲%_3 (:"eZm֜pJqx1^ m5>^mchGNp^0Y+Z19&aUv)ɕ`U$b~fU/wKvvvhXa*,zW8f1 Nge=9pH*"$YpMh-X5/AmD A9'Dy3A3!hC Rqkm#IB~i`v󲻍A^)Y4&iEFIHJJޤ|,bVVF9qpDHӮtC/:4v6G j{Fw 34!N'*M+:::ʏM{4'@CYB\b:LV0Fٵ}I\@d(3{bP}ס7t&Zo&tSy6Ks.q| n qB>M륉- fɗUvsvs&\OyȺgwwpޣSOl3J9=NK=vm 2aV($ ͑!nEP @5$$$zdUSZc|{"0/Uwr;=N2yЙ7n4o2q.N>,,1.N`,$C+BaaœY`Apep@Mu(7{ _gOJP̮T lœ͸>iO]2,\~I#$KI_u\77scV:+H. $b\EJfj]ѹj&&8a.x!GBDX 5̚&c΋ gOW57;6cVŎrA.Փ]XyP4͵Z ;NFO=6t7)ݩ6ȭC)ѡ;nх[=yÀ>e]hF **'9|viM3?\BUl~#}ZaڽW̳=a-Iz96WRɫwSq2sD\1A&ܔ_E_^$8!Fo1D_­Y>ceUnÃu/tNI>!u^ oewly=#だAB@*1iW;4^~yfcJV0:{t}*nηze)]æ^ʩtU+K[{^Y,’(aLz& 3.eչA"z<頝U!t0ED/=sȱ2@|#"E ~$؍L09RXtpvh!Z xgan;9:YQTTm l3<:4N<}\a/W޽Cl4 SNI*i L 5 v$IMԖc,zFU !9tEzύUH`c)qbD2K6* gg܎RN BQ  W _ߎٛ{&BOO|l@v0n/pdmv'o- 8JTBf`.[r i*=ʔ&W*:,ql9GnI%wEMP &Qn"72WIn4 K\v\- {mYvE[qVsF)J`~Rʂִ2_}i1'`zVeY$ʆ% CE{4{`Blևņ6~0RQz$TSJ+`gR)D@ &8)X"NL#3bTi~52JX\1`XLiR#xu09=,JMa_~]Eo]( 6 +ޖUNG5xzpâ(#h#R%%9wEفP2 A`8*)ƚW3jߵ3&.Dd448΃ؙQVF uC5w7ro:Ե`3B}M=jYj# **?YNclQ:[S)頴v)p_INkQkPZ+^OWɢ35Sl a], qݏTz"*#XܵGa" TKR`!E r5s FTFyxUFKPz. i F~vkL1G/xzzC\o9iWd&j-߯/gmo4:f{q~h0KM=zj y_>azyQ?nfr-=CK{U.]-seVìy:\cB; őPGky50u,}WCA=)7n^P-M=>*U/wN^nxtS[?pbt, /0#t8q6|,oq[K(Hy+/6-ןtL^ʚ{,NDŽ=cJ̖IG~Ss~'Pi)XA@'}=}:QԤ8x =8)l y7Ko11a[&]uɑRÆbmw6JO"$aO9&*E/WOݹ4y(6.mYtRRkR,R89Ǟ N]2~˩5h.Vp(~?%׆)qNq<í(){E3(QP/ߒ,=9sa='&KrՋ^%dJ=>XX8DL!M@3D fV,'.LSۢvaz{߾ B\ F9|up\Blܻ^o0ɍϮvٽv8vW4JG>S<<enw/KkJmEԑr@Wٻ޶ndW~ٻ=9 ,IlnYr%Œbآ%e&!>3yǘ=i=K~Hp ƾb F\.nh ql㔐h1E)>D*2$:d6SS\VLG GEׯ_hUJp9t69{}U4aKBJMjN菝wݙܞ݌k텷7ٷ13WKz݋rs7 "\jn-kK۟n/BִYےt1Q8yN>ng=k{oz9iֶVg/յ Q>\3 K\> :7#_= &PI7N:wJI?Y 8j A&, A$$ 'G4Shfݡ#3 ݴ}$' frJ=GqEx8_М8-&{1Tc+&gM=nQӘ*m +Nj;_q9X#q'~]{!ќ}QMJ%‹f<`7AKkle` VVӗQ@4i9zXaX/j٭^>,S6~ɕYF7nR Ric4?BK0DD;wJhH9wF5Ɔ,&BNGb ig_1{UUQ֮e, V@>3еVu$e8Wk&Wx(6VW,gLoQOd?kQ~rEyJUaCJ (MI r$=42(iQ~Jo6^ Sc֨#1mڿLdžq6'T[al' c!멖+\4Eo7.R&%rW %+c Gtu#w5^H \m#߳xFhj_N+BZ2./n`q4\`BUuK8TBDlUhoL5*JQIXj# ?-XP2h֌I8Z!yb Kpl{q `0N2.מ|Ur Ʀhm vE,dYRIgY򦇗a& {%thdF ZOdI>Y 5%K,Rg>?ȗ0. 9G+P&H%ۍٗABT054K`9.ShSED0L)a+ʈN-`A44P/_AOoJ0iZ0u&̑L0?Dpe^6:4~˸~~2%O?_w%ē*%!)Rm4ē1IC";3m5oْPA!Vږ%t#ybT8)q^=ƢM ^_^jVx%qֱ$)O0ժǪчj}0.N p4jr, #$s&Gp [Ü5 :%־oF2`7+9Qgo0}mlx>\fqH5Dh(h)R[ӀYMLϗգC0kw~`馋F{L&cDDdb@A1o3THY"!(Oڄ [F2U,Ml-\O(qO @ WD{p)׮cI45kpV\k/"4z)eQmX.N E2>4e(ʍf\r NkY;E4^ [P۷$ޗ[XYb1eʔp.(&{|OO. nG~wOnhl7>p^$o- $*t -0Y-9B)AZAz߳ I0dFs6^ D#`& x NJugv~X0.OuX kmUvEq hgHNr?XfPk_YBB -=+)lы$΂%%"C*ЊPs%UuiSUއZM}?Br,'#᠇SoqN\\)VFIFhKihW#*ቒJf Q& Nq&8.D@bF:7p<W6GSH6ue;ŤG:Fne&MB5.@\YNb:OGS)A" sջк){ '6R_dne@`.,ẛRm#ٮF)85Me ^1OYAf @ݸ1H-5Z2=U4zc0&DugGK@ AGA@tbJYülLV;Jzk~8x؃"O-»?܉]'7vm4wm $@a h6@S4)ΗYr%9Hd,۔%; 8֌H{smrsIj]}nZ_t*B3fkytw={}Н Y{v}o659gؼu{˛o.%ݠ;׏$!]zigdodm7ܒmg-k婗 ǟݺ}{T|"|wDn%6sYl!=Ok1(ճ_~|EYX'M[bje5Iʌg1V)՘*bS,K˫JȹfE>]ǝEY']eI%Gu7!~o^ݥܣ#1G Cl%0} VIC$:ӑuO>OtKeHH1gh=7WZ*åK`y!uN#D+LI^4 YcA RSfqV4cu;^_|%RaexPQC]ʝW `WTh{86I`Eꠐ"uPA:(REH"u@ 7moF!y:(QM:(REH"uU@.bAɽH4'f oiEl "6P@(b "6P/,bEl "6Pl(bEl X@(b/>FxS/n/?6:" Q!4!Z!Jfd!,EHCi" Q!4D(EHCq5WHCȴD%2-EHCi" Q!4D(EHC,6 }q”'1 X35ʥ}k-Gwm#ZtDDwlKJwFTjhK#:rCHUX[8eFY񊇔+'d#ap)1)qZFe)A@0DF>wdldϽٗ~N H%=W=?^\|] __Л.͋ވk zޔ1*8R (LI 1pil1b{[CNX B8FE'Zɍ^ywۨrBekm yu/c_-D b̩ H U̒B2~ZIPib)Ʋ cx7<7;ƅ; #\P2y1' egn·+o5@rx&H9\?{wv. M̚噲09M(ڞ!x3QÖ~)}ԞEǸOqauZ05sHז[]GәAbeP%>j Vkx7[wpsGz 4&&VK!Y3f4j|sWaFk.0Z{}/oCYWNà}is D8oC'7:wma04\:Gi>c Ww螓PG P߇izt=i^ӡrImnq̅$Wy[2Ĝ\$fxT`T(U3xm2x_ner΂e1Ǽшc8`Ĺ0X)VP h8yP0vY4a#A2rVk/嘋Jx->c97_]zWoJukVfn|FIKP}ж?SSqr,hGyGXKÒ"gHm5 pNڡ>IҳdI-1_Ir+<]0}}&Xoqk/T =N(cNҼGI7>]O07ݷ~1vd09`d~y9`,b7nnFktuۋ|EY? ORqDMH9<@pwy˶FT.],,Y swY ;6blk|6ь剩cHo:{vTzSn5={sg072l%noZ\hY'/^<ݘV^̷?qztk:f tMh-VhumIg|ؚsf{㖛͎svtF,l {fۖRarץol6+c'/TZS#r":QԸ8lx3-dX rpu`؆ g YK&H ,1Ƙ8^\("Darq/ ]LYf{gM{xeuIQ=˜OF/E KrO虪EZHm #]!8u: Co?_nlʀF)h,H"L9I׎i -|*܇!*q,h.`a(!c1Bd#FwN{Cⱎ&,H@ c>s8yِOg< $̰4N+$XD0`AP,@2qs=NwG$ RzAaX7Ո["FVj*FFyoˈb$22w2 O#;D"^ , lJtkn8!z!5;-szK˜:Iw|.0r1ؠtHDsiy/I5(F!QOR=5C)ܸ(ߧ.̮4kO7{{yGZ &b@ZB#L9;9-՗}PCmŤa |ujKPGMΚjHw5 чQЃL>:z|Yg7Yp\wJZgUk*BԼq\3Is>\*b +ViE>ppŏB9^S1۴`zrSz7^&}oc_}€$Dq&zt\W?ݢjjoV5UlyΧn;|zUGQd^o|nm ?.>;{㩈ldzIN&DŽϾ3?k"V{U aU7.DZ~vnkc҇=҇V1[S0 OSlr}]~ܵ5,2V;GJQ`x޹\?2+rR1~Fbe)$KD6ʳ:!)QDEmi0֔Gfn}\#wIa G"G!T ϩ!JY_ Y-G r96Tm6&Mtw݌zz?2A%j7WO5w3'ߧl &cՓy?cgmI L,-9R&Bz4D~2d퇴ӢrkTEg3A{q쥌Ot1{'0.qůJ  Iԟ&92کLjJѢ:b2ѧ:'%нɥ¸oT.ؐGE1R1 Dn |ޝN +7b~sVA-m=3{9A 3tfk&_~\ge[w=um4g_w\7X0Sb ^^r:gƝT]m|M8j۬A̅ %R m 5FzdM:\P`l)%yÒATۀ # XMN L=T'VjA8zvܜ[k= $ y*w,_ZV w6VoRzupJ:Jw20B)B3XRz)7*v3=MịUh4$ 62wF`) NP}tdl͹قYOG$v_{UB5!dGLĒkY(0眜]3DDzb5LS[Md!HrKrPn(#!eٔX"B.agw&'I%ōˠ}Aj| A'Ffq쵲2T5 +B01ho!4U>=Ar4 ]JЖfG! O4+qDiӿGz4qΐkx Y3Mm|uTD3SҔl =s訢sVHk[EC\@"BJnQ1G EέRȰ, ƸKkZ )$1vQ CKl%Ft\G{:p8$a=܄a}y > Ĕ*S*Y?x;< ҁ9L;-p2R_l\o<РTLS:T*Qa|S? -OԼ} 7<_6r5+/Nz\ծήVeb4Om*^3]Li5+wm.RySGÏD=X%ˍGV_2pC5Knͧ| ~61U~Ϗ1YYW`"}U`Z}}pt`0 3gHm0<" pQӯLmw/KW^C$W1tHc,:RkHZ+(x{ݻP v!gd }~P5?+kn#GmGoIGÌ0=;}D&$}&EYd%J.mQ$‘D//G,}8yI\1q[sQ9XZnT1_߼ >-.!U0[XeY*j4 vL򀬇vOR@;) cAI+4%+Fx1G-D#ɸp3zfIw׾W'Pwdz+ P8Hsޫ{uzΙzuւ&Q'w^J" P@# |yNL)^Y(AEơr#3x{{K%`u<u;˝}QLNsƋ:eVKh~8vy<߫]!ˢծ:\@-CrL]}O=u#w=.>qZ mE։U]^ߛ]+ [Wغn]-[s3BRe`~&ebu`//ojf|ptl;SўfD+ແW++ņ̹B_eƫP^YMp̩GCY8wt~<1M6J;IEVKC%Th, @K'U3W_yYEU ^]vw~(w~hw~0wx!D$ƒbQ)˒!Grpdr' rD*~2lGm:'[gzNiī9R19K90 &pS`oS?M-Bt3Wgrj#~tS&jdAYs% OG|'Ж2kj]; }oӔ "4(eX ڢܪQIvUw[B|e([F̈́rwD'@Qf+-Մ lɨHf#DRF!R)Xl4NC0-"2:$'\KC`RNGPTqG.!HFb9Gbq,X({,<)޺0}-G;q_6h49b[k|)8GD&z擵:ÒȄ0| I2d#{YMtP+Y =blbFaU 1MËR9~2 \Puڪ0j{ vb4 3Fc /998Ⰸ#+G@Hh9IPe &tf,{RVtQ#@$A֣BD7$ ]zEņsk~ lRq,"ˆ{Dq'Je"AMFuF3JEJOWeKJH6#RpjҔ,eQEE5&iCUt)Rp:V{*99:Qɱh qŽIpw1h.v&QRnL1.ɜNO炇Ũ\'1˕/*|pHD_U>iżލc^\M DFѐd=Dr؊zCV&>jLJf(93Jb@1Fm`DT );#D=ͧb~xr|fkB-a;O)-?"w2WJZB5 :$qƱJjգAbJ1fdYk]phuπ7,\OraϜ)=mpfخhĶBS#N؃[h:Fy[0k1ԁ UIuJ֘Q4h ag!BoV9Gïc4ueRUE6߉w0, m}jBp9lDW,%IL@FBAL ?7=0<=`4kE"`s"H-%&J\$R"Er A#=EyO?QCMrw{[:&Wkl(0jBR  4 Ea~JIKfF#cȴU+e%!lWM3Q:U$nqVw# bi9}# 7)ߊI;;6*}~f"laZACjW,PӪ[=w םaƼM9 p$ПA+y`ŝi2AпzX~`=QYQPDKAHT'`iG.sԛj3W Ky B}Ze嗟 oƩWϬ2hV nߋD͌yGA4@$e*i/S6j>q H?Ft;).1"@nz'xZw4`?- M&*6I KOz(:qvOZMmLgqf|y Ҫ銷vϵMu{5Rŗ;YZ.ם߁=(k 9jz%8}*m=j.} ^1/Np|0;EHц]Z[bL zᗟ6d|\._#D_^,y`C]+.NW;(oCX47q5U}l$qn;QrJRۓqU5vj!R|krE$ 2L{kxŒ5v龅 >ϣ *Z NK ,ыw$Nm ؚtQ S;IgHxSs%F.!wyHMP2aa? N₨ӎrAw=_&\R>](#|`ܘeL N$| w>/nv»C]-_ \f9]}P0K*kRD|eB zԲ'K9īǎW+ϥ@ȑci\hϸ9Z 4p¬hvCSb2eAqʩ?bh4,Q>piԚ)cиp{c89QAdjʔmiKtbprW ݸ*G2  \ opG,wy\AGcp{sCh*1pA$,$R;g H vD!HPG@,74(3 2hɨ Rl8wezq8/f?8Do=E};:I K_M@l$­Ɛ(`̒>ź6%yLsޗPRW4hPLFC`N6WF8hfzBdSQ>BoO'[IFT19FTJV|۷3scٲ?{3G$!cuRK@},PwO.t?muLɟξ)dTrJbTBrQ -pR+eJuY "ODH mTIn: Z56MոPwk_vWGP$K|eaeZE n>o޼O>[tvxt"Ъ|t2zmEioOVNu"%+'=5oXA@X#[:bKm(TJ' XTV;\9fRgK;.I*Zfm##G88hIzN=0\(+ʅyfO￾r?iq*'}_+ΨA27rήٹVSd/l2E#xy=7DkY DY*|htv{mw'>6b?d).{UA vxӻr!o/_^^6(fT&mREV8~k.!mR9j%G1Jym=}zz1283_t}Ӡ~-sQ6zں2p dMK- yKa뛑YwFy% 7ܶٿO7 Nڵ2 ׵v1hu7\GrqU_O~^Ahoԝ|crnTL?BO݇o~m?~eÛ(pvyvA݅E7Mϛ_i^ilo47beQ&|vM!i钬kk@r~OȞ+|{nW=7j~٥OzaVE۫J^*\vO bumndclHh~ww1[S(d$lgs8jڃoú?'Pd4Qp @2rc7y4S9m$ 8ij=zAXq8WVq b-YD&:xGw:NknQ{qdlaYS[ZFWط\Y!c@*E0 jvaU8WEYeaUu-O t^\֣ -͟VkT||}75MU gX mT>"z!b!X!/R>ID3j ZX$QŌKIX<)<'69ftS~@ehUq zNHSV2ATXwv(gHbx.=1v9&g>-ű3>|4- q'Z `: b*^}$(e[pҁ7_v<,q*eM8H$ì2 ]VzP50fI6!S uD@JG-"4b)\J8.p"$IlT%2d/OA34źnV,3~Նg&sly.Bř(ַr|TH3q.\tw6ߦ9x96˗%<"ORؠ͵@8ci ?+D'8qjFH`'spYPrQiP!/ ?AA0A0-d1qjq~q5uaݞvx f иEWf2HFq6n1 : LP)De H7G!ɕJKe;.rL28lB ̤֌ɠCJ\;5^ x<MӺų'_=n`k8E(7,Y bLK.xߡBF[ur3&wvmΖ-UVݫ]M:qNΝصIΏް\ 6Wkh&g Rxk=,`vtӕ\=NpEtgK6XInoj-l%qCK%7C>Z_s1"a*Rxp#Ψ7܊Ew}jLZΫWS\}Un8\|n[y)/4{ r{[L1'ˣM"5"xmd>u4{/H }w> uu~+8 XYëȖBLWBDι.ƪR7L]OdSmȁt Z]Ls(%0fTv%FS\Q:aA%95'l' 7FD˜1Xh'`,(ip4xG-D#ɸPK;a1'C4SSxѸyJā`|m)39ՐMw$)AFEh *P"ʞhGAEozG!EIټ F[lk"(N&z2Z͒ר0HFN'Ӑɞ$3βJ%c6 RmJ;!vGCXbIC*lUn)F 8cD:S1ck8!x} $Ί%s!Ȣm D (yHڕN"\;aK0 -%E"VIJRvx,8b ΔHSՁQDLq8*=^PB2*JMJSZ**y4! fBs[ٝE,֝w."jvq2EqJ{Ťvj7̂gj{۸_@-@pѓރ૭F\KNs~,^[V0(HZrə3Ùy`Ӊ:O8rldB#Fr՞6 ]}IǾC=܁ 됰Yl(⚻J= :~}.S+ռ p0V LPR989 R[ϵ J |4Ti&X! maS6Z EѰ ?" Dh Z" N u9r_ǸxE+;buo{c|x)6FqR4aJ[D92$'DV&eyfW 3 @7b&3Gj(Qd}H h\`ݟ3$h 2%JW,s~RҏcDtSF'q?N?۬Gi-vȆ)NwO[$gC(e3.4 20tQ>pM8\Im{ h}gZX/JߢТEg] ԃ@Qמ]qw#ԅ <# n93a$EL}11JQ)FJe)#hσī:V˔_h=N_cw7eWa]Y'Gzj b"=ïtmxߑMgp~8UQwiZh=l]=E=a=>11Me{uGC31m~ăe[0Gr'__n/GƉ% Y$@H븋Vl *+tx/'l_S}͎{lP|g[GMLŖKQ_3mGd;ljaD蝍h4|"ubtR|_xN:q-.{[2k{J=si4ݟ=?EQMQۅ EeT"GG_h+)TBHjYd6+cþu4g\o&S-5AT'8@8pyEdV0t%׈L&2h3dݞF\dC2FiYm\D"F2A)&;k"̭xve#X= ǖtMxuW(hAl4qJ3sJlX)%u [_ᴦ(ku ^:-!46|\t:!蠵^tׇ v֝*H=U>+x~q-$'xC$@@c Յyb~F+Iya)ʲ eA5G0+~G(߁KW> zϬ^%겖zlqGcD58 d ]ޥ>I>AF] &a#uI|pܵ{[p{&f; ,#fz_j]iaAt)B>R gDfgNE]nDv@{s:`Vڄ`ߣ0 uJ&Y(A #PRjzFLQS6*7ZUiK(ruRW؛rj+bo0ʫ@}:ˋ'\B!akoWoӰs}[Yo=,M2\r^ ^`',o3I@!aNkf^-]fw"Qiv+{ _ {"rΠ~sa'/*x'ʊRUu@D墶C_Tx L?A [M\D)]RYE urA@ 8GZS 1l.*`/ hD^|2 Y]κkɊAz!G,WԔ|#3l .7w>uEIs[IcAGO8׻PCr@ccWyi*%RC+S1-yi /m%U:C>vx|yF9)71*Ytg!~0@ZW6hϛy{{ Zee1 4*SH)Ѐ>f⁅$#<B7jMp4 CH/:~1?fjŠ@|e@!#C@S>$j 6 ъR-*B!(qEM:E<#4%ve]:CC@+sYR zQi4JKBY*M%Dsi!/ٰ KP;h"J KbDQs; cp.aS1"uHv5K;L uY6?Y;75h] 0ޣ7l9"r'ѧ/~ShIz_i Y!ߔa|n}U{4n\fWSB u~=Nq{+dqi! aPN99:?t=OuCڐLG Gb`իRbZWQ8k݅MΞ۫)m+E׊S\B5%;M50#~u6W~/|{99~fA#94蟜6c.PԆ?_NT7v8ABZZښa͈f*bQmp(|4\z|f?im[d[V9g:(>]1گ#zH{fJߛ8㷟}o}{۷)oߢG`}jܺ nI]5͚iZt_]+{h6ovh~ ϗ㦲g _MU~]\C/~F5?㧏{^iB yYMF h ͻqbY1`U̖T5JGS?6 ꍰC~ܶ5UGLj:$xbD Zsq5\aǧ?FľoJ_xWB1YE萸h{8Z“#Ix#3 ]}{`0(ȉI'Esto$nP#Ղġm4ĈދXن+dbJ7ȁt3e!^tNɸP|з'CS:lBH࡚\HK7A ꝫ2J@f +JZ˨< h? <-q w^ ׫2*ƱݦY4ػet)Υ@'T1 $N.2ڹL"U` 6L4 zQ ˂4ճq *(k].+ktgۂiMNVmfWꨞo+Ɠ?;UK3|U>洲J?щTSܔj!O!5S{F !Qu ˜hϸ$:ʨdڦޞ$t|I@ZX(b@7hdqq,<҂F0κDc_<5DžY]|wzKȇ7[T}lƇ!MEy̅xƧ0-.WīlMFxuzx2/!R}w59y^-Fh~}0@Lil[h&A]agqۖV)SZ2+ \\P`l)eNJ pDqܸH0%r毗 WKcvS :]Ms#9r+:ڎثc7— c_l!$MSd'b)jɦ@RWDE@0QL96)c6 N1}:K @"f:oV+RltMCwj]_eNQZ !I׌]dI[t(SH(Ze/*%BrJ ?06,0CQzMN3K6f8t2kV Bp+fF̲w+>w&y$&y4K%[8 D~v;GcMy'qnN*7.R_\NWA{-l1)@8S̩P8M%J*4`1za42#[à)25j17[w>|J4 0ʵ (LU504=xpm4L- maV) uT0f2 `gz J($"|d=clY֡(@HXikb3bmŇPtgM51BRQ1zQyj}L݀-tf?X:~YWk8ܺ48an~Z|iSP.7}0o2v t&V9)W9Xk3aPɁ٫Ga`aqγ4B2B[B}BCv轵I{L :%Q)’T 5{7V+gQaQ(ɔ*h$8@/tuS6{%>'L}Þλ(A2QBmraJ^HD/Bd% Dy탐h.&;cЈ ^Ck{oވH<u8¬֌!oѤ:;"vDK"0װ*VOjX Q*.0Eh&5S*Ƨ,L֡z d0TZ7?T077~i !R̼ہV7w3Mx:yAmECzX}qԥMw7ru3uI}{-)䧯Hv$Ձ31I ?CP`8$"j5IdT9=#M'9$iAʢұ-&XY_QFdEeJ[2L轀XQ[Qt2~7c}׻7_lRiMWͧ{ i}3QK%=Sl5{뿿n"=ZZ݇~$yv޸2 z2eJVЈPX-ZFXEɩTBT>N`a#26gelUf3P[hB;£G_*29}8q]smic6Fb1%Nϐ#Jb W=+adT{_ !1ɍQiUͦv>9!N@(ʄl+)<>h%n'czF1Ϧcqv`"4j/PR"1y;(fzh4mvbekϖ4̢Ξm e (~c--X !{l8ԯ|a<j~j}c8Xm3,Uyaɪ"ʵgvg*ͬJ2bҌ2G6_=H!*.[e(KMިRLd:CJ̤=ʴWJ=y3?vq-:͒Cbhla]6PkdJݰ"cZ 5(cMub1)]<]<{lv= #$oy];w1{4i'i8:1DZdďǹ&wΘhTK<(;rP"{gCQyp2+ 4UDʶESHQd I0#Lвt!CiHoEy\{t_,V?^_1f&D+Q!=O@FzJPDU֊WjzX joF_fMr>B}ga)2'vm a], oqWǦؑ{߮[]KRU& @'3,Q!Do`3=WfS@aB BS-4d$o85D-N2~k!ҵ>2,yz u'!6y^]^??Tן\[wz1ں{h{َNl2n.:\]Oww^(Y֭ouu{=ڧ13 jXSe;:^,y8{Y/F^3i\ tw%cL[osҤwsG{sqsn_Vt_%}^]XJ@6ҙ(9ǰHEl)/z4F8cF{|Gd/Z =PFLasg7P!,SG*&/\2s\ ,Jѕ6WO:.&mAY%[ m<8#d"BfGYya5C ce3q$ُ\x]9?R&u,|>[@o^գVRDh)]؜dx%RӱX,Fd SEPAs.B,E H6yZB5Gdٷq:/e?d Yk0&AhpIk*.* Dua|[9Ҝwv5w:N&|"ONYdJ }J:ʹR!It\Sn'w7yI)ZMCBk*GjZߟ=݀{3[g7ژ*}&aLTZt`1%+ْkmeiJg]1tEQ0ּGߪi?|6fϗG?__|/VtN_ ˿O揷q g7m9E쌷2Dhs xU'c0Ȭc CuiMLڤ:0;ӅzUk(@j <|u/2b4]_Ŝ b|?r{ǟǓq樲θ 8 }@B5*7 _NQۘ?1nzXG8aŇ= ?@zN񋺽|MRgռ4xڣ<1v1v cTVmj֚M&$]$&#:/EtbRnEAkd ,#2  YHEW)@N#dRGF)+[Wh%n}]nݺۉcTq U2Im;0cv.:z#t@z{'g4_\ 4H:1v&Ah]WKRvIcŒfoE(O)vtU)V._) Qv`E3Nc@0Pn[5kX XWe= CU{u?+f1 s퓷jo Cc2OqJ< G%:͈hDZu./(>gL) $  R@`bJ! \Lնϛ;j)`\GuH>]<. &W#%/xoNlxR(Eht @rH%"CQlKAh\γaPz0HHF: Ai/uZp`hP?Z3[UlZFw>׌PRzBJ-sReSѹ`a{Niدj|)~Ęf3d=Zar`vsMt2M2_]-Ud]ތdz[Y,+#oc?y^84QRz|rwW=/]W Qc??_f~%~JۧYxjl?&v8tPndeeYy~y8U\H.iOQSa1QkIL1UIEZږM/`06ͷ8`!o;|U:=r_GmV ? o{,{uZ}<?Rъs#+ zHPLےSe|3Y:-7fx",95fG^1ڦ!>HWnzu 6e#LsJGzθ {a,PW9o{l^us~3[hA-C0~M'ފpBvL]RtdDuuȸ'$DcckZMfݸ*ϻ1;3x3Ώٻ6k Е=g EM$S(bg^cTTl!Q)YIlYcϜ9NL0=w)*Fi| !$D^2G̥1^̂PAJ+QӦd&ww['[=+U/Ei24!V2&/ N͆848~q^KBLk!~JIdq0*qGcahWVq EX2jfQq}ٸ8ոC4maJN$%u+/;]4/z|cu)-tOxםŕvAkyu0h%b|ͬ-EHh$AD,-PepJsK+"!-D;ѹ/]:Oti&6/ۈ?:=ZS e[y4%sHò)g@4G^D2gVYx`"ӎT1z$6FgDڂ73C $Wi¦ER݅t'SUwfZX% &~~~҉+>gQ\Z\1є4L>tg)iTx M{WL;eg#(;g5w=9?ɺ=Ի29X>"ńg\2˰mN|]"gFśc&wƯcZ2mlI[:Ԍ܌eayK 1Q4s~|9EÓlت`[]rS Bնr5'Xu֒ʃb<|0[?c塚oy!a OIo}wߖ~}orow_f`B) ]׽:`=:Yx4-k5M-nӴ9ńߦ]9 ^|@0nInǯa;/ٺZl5b$V<,qW8u]+`)U8 wբk9V&, UݣUVT3yO+{lQK%}}!)?ds.|Ib`NG u⫍SlIo^:XP1HrXT2+$D&gVeB={󑥆niHvtt$Yot)H^t$I9iEe%[4: x:f:f&V jzե/]bA_X~_q%TWS.ĺ/f_beM-qG?Xaþ|Gc'F ijE0ڹ͗ѵʄ[R=2]=?Mƈvk\D)+A1gU:SrJ@e]R$2(- *sFY/&DH]C<nF*w F_:!9RGt>*T:B`)+[:u6:Kgҥ\1^ݽ% /myE;[ЧV/&3D`2Ủx1mUl++GL'qc?~|_̚lڄxa]H7(ljDD\0vGap,J[M'| KJo`}xøZKq02RJgc,uR A9!!2C13WV-QCs:|<pPڹgm׾n֞wbg=<^ /k8OCWjRJMJH*hZԩ:JFBL15ѲK&8$9) 䘤B"~: ˤX|& 9!V$s`XyW+Zt2Mkv}r ):.WiVp]Ÿ́t)ۨ6A5:F('N&[)L\*Y!e}) @%%9<C}'/.m=aj"Lguf 9"NP cdNxW\0'A:2%(X% :zk䙋KSk\ڞ<rO*vUrDbͧkuCð  )=m `lG `A!\ r07{!-al(ң&)]1k}kZs! Ȳf\VdR:;Yxo鯰2kן֝ifMCX:[hM) 1&:!AR2]k?m~~ě 0J(/63a؃g^(wB=; +.\PZEAkb}*~I'N'/XU,ڽ\RNCK_|ح?. $d9TJ2Sq((XN:!\Q/*w^)Z3l(j}{eP bЯ=y"~c*dx:+68y!VFc}\DQg 7hX +׳X>6j0%vqLRQplHL%.f 8'<ܸʹU-K !mց7-I Xi{at|?;7iだ5/7< 1By(|Qttv4MC{Q(޾ "٨DFņc&Pv͡@q(S_ V>Ąp)1DZ I#,Iʑ@ CUJA>ql2GD׹ASV-=wW~8~>U%-Af\(; m,{( :HTݣuTV@v! "1v|P9 r&d y6s%oiz&HDF蔹Z i(d#\EmJcQިh6cP.)grH sqD1wc-鸫6v`,()( W^KNAB Ĥ '7X!C͹2u3gmlYXL,UXAJyuKXK4kh9NINȵɛ3Vh$7I+]6hEdV&PVn 2-yjOzmk:I]׋^&+u`¨r9}`RJ%LBaUqW^/>^}XM:vEPaI(6'iY:O9災Ht '%' *Zs:ѫqjrThWGU* 6&;k3p,Y]w3+v4Nt6ɉUq$ T*A(B)Y.8$H_ZdcaW^r׍aW6k\cLuI=w[@UldC@cDshP-n T\ GlMQ9vti׶u񖞻>w>pi+&6NF(t)CS}^L{f4L9V  А24}`"/#Jim ai!z/")]mz=@6{ã]tɭ]k'}5XGvAC.osrgZ{8_SX~0$n ov}MT$4eg!E=ȦDI$iOuW5nKM7T,ޖl*Ē5\l_Su͕g-1-j^)~=伕-G{mVA 7ij|CmΒ[4u9˺9_/nNZl3_,3zI˭͟6rqۼik>NN6;X,S-R[v!5EfEkhђEc[W 7ӜuYh:J`kR&mAq޸H NȘ=+˃C\&hRHB 8QȤոL%celid/zeܒqsi\N||\<57?ZVSO!8E&d p',ŔXwAe#23zLxBh )hcL=I[)`1^z\2O5p=:%, KtQԐ9!I%?)K(.n6D7^X!,h[a/4'4ͮNH{I¨ R*P"I@% ,Jp_3Sh͙SRZ ddhu}upa-qYs<=Bw:EÖ)F:!g:+pN'SIiHN:['&sD!s:޷fV˪}Oq:]ݪZ] fq΅kb^T~<[_"+]~zvs) *5 RK#w6QSR$ mTIn:)t".5n}8`yz{^yH_hy8:jfy{B DxSU7>Mksjԁ9ru{p5v@/CXW;Q[wutIԑ|L,=. V/ iLa[C'zI4NRR5Ȉ`L-j b9e=V?mjbRD"‹m?l|o-q2m.8ؕhʜY̱w㠹h#-$ٌ߮poDhQnGMf?ͬ-bTBtO*~gv9M:-Oܠ@j fGΉOT _ǗeS/&Ϩ&Ukb~xw&Rs]Nj4Sсkf3ۈr:S#jeP;trON ǴtJF,!.fy۳PA(,g}dC.Z >Ň@dVw i.mӻw nr#s,ٷgmo~ǏvU$_ܦ'Պ[͞g=Zgƴl3Br1_,a~}N~XzyJ`#>J{C(q V^{tV;۷;ֆ^ܤi{}=N.O`z?bi*4{+|a/wF|Q]6/'b7J$w_z\s8R'~Z+@!6D(0% E+COs>-i&`E 4yA @e3 \&bԊ>. I| \/=R9j~uO}]1e|Do0H ]G0 ` @H"VSxB Ƀ}hqm@^qۛΨRpy@jeAN|/ ]^>p~ Y^=Ќnb8%dkġXMt'locDzjc[5Ѭ88D$Ѩ \(, #$RL>pɁ- 4`b IJmT%[ꩣsNL+Mx* 5g30^Sgvu=$g&D{D'Ч!w_6!FiY%>CCɳ(jj44TOG?35C|D$--!(\ 0fc>0BBM .w@mFɀ/ uTL '(ws )K$)PHM9؜<1UkΖkUAśQKWZL]}ru:"wd$#  #&:ϝI)^FEhS"mGAE[R9PQ&*"I > BeИr28sjfA`|zXx3M6<.`u[&g:H@DMKb[VA+-Մ lɨ|oc%OF|`|bQT3*S9=l1!Mx#uHֱ1U(*ҚXs6krX.,2vՅ..<8В#fvqǍasvu<@ j8^ƶ8/!s<'Mt&-Y-9B S(PA0< lA@#9YV\iS"hfNhc6 xmJFjfƣbb.:vڪVZb43F$0E1pp+ $IPe%t,k @+QׄHbG2: ? D PXkf} ǖZjD]X#^#q݊O*4"=^2ĦFuF3JQE'[C f$UQ NmRKYϸQEE5/GfXsp=8tyLgVɮz֋׋^\p>k. @Mι0Lމo $q VzՋǢcчv7}x4zNO53&Q4[8ޏOhScG~K!SXkaVJ_;'. 8BdJlM!Q+i#t!?U[;BJf QLBDh D1j+ [a9MDy2E5/285@Hi-L%P` l 1ȺXst8d !88t^ݬs1|NgC s&9j ոq -A"0"gSS6׌)q/dx&٢ ̯(KQ=od1_oخ*qR{KЮgʚH_ؗ٘E'³`< NX50f'7:$D 4anՙY_VeI񹦞2!k=3W48m+re.efB9S!Y0=tMo}a]2D5EWE74oʦ 9SI@h= \#f[M7;e1Mw瓻݌Vv4obW 77<-y0,M缝3A5\SϪ5WKivGdXi.:ADa6⦅·l7[?oN9#q߼nú 94hVǜYs[>X̣ GmvѢnٖ商UTjv hK#:rCHĐ8Dh/2NـqVh!E l$\.%?~ Hmv [pzts 5W[oέR[ꨟPGrP^t=v38854&հ-m0÷P.m+\nr9ģҔ|6J!۠Ln$Ťx-kV#)"b%yG#6`HZA#+"r 75*Cz x^ 1[,} *4;@SY~uuy_]VMj@-20?(ٛ)Q:MĥaI3$7ڜ 8p'PҤCE ^Y2wg6t]4;Dxp`BQ+|Y"#SaDy}h5Reu)'}[L׵@L؎^[rz)b~g_̵%*2\"sX$F#D*(Ghs{L:[dg@gvNk *` G|d1"zF_/b$͒pQ}mknæ~wgU.NSq5Aqw}ҁڪ cgc4"j1hk,aHa7Lh*$фS g5==tCNaUhX ptNB@G)07.W)b3}_tA8`!5ˆ!BN 8;A/]-(}aqvgsY8\y9-fԥQRv0\pz"[th"8 J&As]Xhb3D]3$[rߗYMz6F0g93}ߛja)SKOdK .nT5 ӫBLB]oPɬ\,<2\W5 u֕F=˚bEwnn]8yD~-?QEݩ0̶s#|?CG`e2wRIry4*,B)@80]ME|@~7}_qh緒?4sɵg']z) 3n]n$'9{{Jw{쇋aB55)Sek.-'A|;)RFIɀ d>sw邚W+@u83'>M~2t+y{'^N Lbe'AiNC?3):U@r]gz'Gtb8~).;f[0 LP S496.X&t691 swfF O 1NFɅ莯6mqͨ1D7Y X!]c9C` E1@u$|?D_qto6yE@UCNe?ET \BQd Õ_m)*yhn&ffQJ;vwyWW%hmJ_10#v܄jMr[|/ٴ;T :޾[ݸ&UVHL)%viўJP'(=d{S8juʍC~R5r&+ *0wHst]rЌyTGaXY ,j|yQ5mC\aգ&`ۋzJ 1@RBJƸVAB":JFژ cLke7&[$99 䘤B"~:EɱL @$s`XقuoY/g"ʸ*U`v#km( wJ&R#wl\#P; k1.[IYF[X{`Amv37ཱུ:roضTHFHR=d|&BJA:Le=dp.+nF)Jaϝ6d=ЏLFպs6< +@g<4<4qhC zI*j|4HS*WhYJp`']̀0פ{=xK/NƩb7K;/Їr9*P J 2N^+X}O%Q'_k']~oz>:o|it~?X3],й7K]j47 //:M> KٕKQG0x*xp+ ^m7{5X+}o~YjoY';5:ux.>1EOVɀĿhʡIW07W=q Y@!wΣ6XJC)6w](yu'5]nV[q, tK'%}5P߼qXC&qI\ˇkp--Ŵ|;7|^UN0䞛s{nrM=7&䞛K{nrm7&䞛s{nrM=7&䞛s{nrM=7Z䞛s{nrM=#"[@{nrM=7& ՝ u&pT{nr e5&Ҫ4Z&rЪ\4&rDA9h"M䠉4&r<(&rhK+׊5&rDA9hTL;OAW'mBNP@DDcIin?(E0p)O.'.y.|_ a\F rV;5ֽ#d:~Zuoӎ6Cbq lNH$ʁae ֝g;M&+سmVl=+|$f5:!*2MGH@PI{u6;x8kgerw{1Ojؗ6j@ë}+N|kY7zOˡVA4n_~alP6!Re iE%:c9qpmٹ$o]H~mE 7PFՏ-7Zo.S1E=G\ҬS>j"hTNX+;wXk.Cx0V)Y/hg$('P6ܤ˺QJHCX7Q[zJ!7idN-س=}BHgm my$#wî0.g_{p<kcD@zdr5*v2frFciԢ4:X=Ԡ}d R cd  ABI#,b]LV[LbhXv'Wj~ d ! [MwzXK:S03Uwi 9sC]H:[rY)&< :zfCeC֝Æf='XJFHSЙ6Nn3}^GgIE֋UW>9Mqe֔ |b".gGgXT̕/2X1fHE\O 6X 1;/"hsV)n$rfrR=cJ5_XmdʾP7_xp#PqQ#c_aUhm_h4</c;g}P4!C-<z+LVѕ俺VfE=DrR) Q1P.)gr()`kFa'xb>]mtk^4^b%`k BFkw =ey1dbz9WWY(K˚!WĢL@&&Vj{4d 1 m]FZwÞԯcc|28GGlq߈&K d" s93@4k'$ 'EMU,gHnVlJg|d P [LeX;G'z՞|:EW/_[<#ZHQ`g3/ߤt\ZIhK!xUŬ6/އ_|*x*ރ XYl }Ap-5B/:m %;tZ-j(qI@ʝ#o)t $- e?nf;1#)VɫTP\2s 'δ6μ+9j2jűnZS>Z\by>bO; QƠB&:bm@o}Z9^22edԧl=01U٭Y|]=$#%:z$tddw4M0DY@!SBmNSti`ZW?J$w|\pq>SN?(k>J?m,.Ͷe5a ?gpX8 a$ J9&䗳6d$l QcYߤrVR9-c/1Y6~zֺ;???`APXI%FyVNEB9&foT`1 #܏G(cy5_@cP<T6`+"w٤uʙEr>ı,{p\\ ͛? Y~t0/rNqN]̊Uҟ&K0Ɏ!/Pdѧ,q:N1Έj-[z?a1NKl4b@T[*ĉF 믷_TKf.J{!t z6xe`W 6q˾?wz {29}:-|>ӟVMS`x)S܃IZ 3 89RK\NC^{eK.4pp~f7JlXM}pA, @AaI_1M9MiLwMw|{']S|Sޖ4{?K~k?~<놧|نe({Y`PD-H:yINejW8@ob`-<umbnyts弌@3.t*sdq4 kxU"pay0=d͇Gh=? {w< E-'|9ߪ"T]3w+>lR5;k# յ*~R߬D9},s,x>`Aũk G]wn7aʢ7[6z 廱x/6 r4z5X_OheBW[<^0^v`/K+du zv=[O\,{:0C nLR~8:2@(ah7Ō4ҋ~7,,><޼3B!b55 שnx.j=oݺflIeU` (rwm sI s'N]oX,"M!7AP@<*D6eH'kI sb16"Քf&͂|Vz(!M:[<7ziqUj8JH3؟ &TZ6+uݑ5ȝTvQ{,Z;ac-"K0 WG;yg7O2꩔{ -?眍T5D+eԄ(ZS+H9++ZVLӦ $G\Ezai䘒1DҀ%Y4ҥPu2r&PNz$' ahŀisJj &Kkc0GXH&=}UيW%wN0TBUOU=_MSߞd/W%pϪ]EtRy^[_}U=/<}O/ %/_V/ҧplJnv}7۬+Yv~bby <۵=v'8aݱs1`~蛮鋧^|[ nN\Hw.{H: Qr_,=3}oUýDuZOcN8iEƌ,a;SǪyC*Kf)}ˣʅ-g.ٱ7旝ugWð}0!mvp< n|,yD1k©x^S;@I(sBvdbr5M:G;"JR'88Co3|hi,Hi=twaݻ}VnSE d}ë.;tj$rsL? &J)JetEqw Ns<-cR7Jh%mb'fu&>Xa*" T hHDdLL[(4L;P\h̚ 3xeR1Ak h[ʴhI:!dly*c%sIΕ:E4|)iR!%3F`)^KA`h$t%w8Am y'5>r"Nj5Ju^ɸD|0ѣlY.HV)@OUus0dL߹q*Q6(o,ɢh,xƄ:.ujXĽ0*2dn< <ۗCxۆףۆch޲6Ro c0ĝv1)U>1!S\GK G+eA 2 KtCΙ$ bXg!d *31hL1jb S]ro=G&K',Q_M#-,C\Xߍ/m Igay ayk yuREhAFFC`N lqĒ(qOSD$$qT;L 9p6y߲݅<08"y6;j(Ջ7qV3ܼ _SI)Y+lm{ 6skᬪ.A.rLNoś_I0( ncK}H/DB X>v,NY)QjD-c䵰2]gJ}6fiL'%Ph)or[, Q^{-)ƫۤ__sSGF,KIZ4ZQ*#K:W{^rlS(z}we0uf=)'Ԧ] /J}81#u1%wpZNϷ"2Y@>/Ii;r_oE-#<َJv1goLbjsKr/?x^^. L6|ypﺄM/m?wWߨGS͕7VisJ#WkZMݵy #PN}oZNvzoP(cٺu^}W`i!*~mdyC !{ttjŭRB5gtmՄ3ATϳ :`G6xrcL؛_lx̭$ҷ$i_sNAϵ]⮺TXq}0npF TQ\@\Py\9|0sdsC۴wL ;W-DJ#۴ /x&8gDI[JMKYGU _Gjٿd{ ȞP{f ݧ4]=i W8ŲKV$Br` W+/#a`֦A7k˼I^S !us"m+ Q>9 #Y\x>D-c]bII̓HfZK  1PR,q&R!eF,'ye,wVz=Sbh 2rXc!zƕΒPJ,HJ&-LTDH2 `wt QI313(3ˠ,ƆrNPD sDU wp43 +G[Iq4j04B}}WP24nI&9F})-a$.AV0ѐ*]lNh>+rac*p!ނ;F@lNJ؝Jzʂij(kZҖB{aJa'Z]Rn7GIE$,.z$hḽ!Y)(lw`܂F/fJ},JR.qL28 00 Fas Ɂ1j-\KɴS;E0B|Rx34` #Dʁ6n.smk."Qb `ґf` % <`]vB$Aq..9,#HX=l #bsd@Q6l}"ZSUƥ lxGF2a דN%PxU*_BM R=! ڐLU Al!ڠ . 4$)8Ya,G>T^ 4CP/F%R(M˳ 0=) X¢B*b:fXRN^t@RXce)'-jd2rؕ%I f 4X .jv3X(+#c(<(ᣁO:dDr9RV <|oC;o>(p s8n ȀٵC]HC܃n8Мګ\D9D>k wzT*]@z[[@voجWɰѸQ!|o/yE \̽sp \n{)I^k"<2apa|+DDp # YiJF\TŴ?U ϠsP\cGT "&X2csJ62bZ]Xf=9K%GEk&p D21ֆB߳AْN)S}]xIk\. ѫ`2 `)VXae%f@(Dҋ.ĺLtx/K8>(5@=Ee(!etq›TV̯:^@\0 5Z>%bU(._1Dd ¬BЈHb Kq)z,XcbU=\UuuEC:m "BpmY5P GuW-qZfZ0-wrJZ@+(Bs,CL7ANm-h=yrP bAlAm٬D]mvݸ;߫Q()OI T:J +QZVAH ($G"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R}J 0R`ՓQUsQUk FwWU+&%зRW@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJoW &SR(`qOF ZWJH -*ZcwPR@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)^%Z2(`n+ǯ2°@ߢ.7H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@ߐbi~Hji-5nq ;QN06.Ǔjddf::UBNuu#?#gmK3X^'{*X .ѴHk s+!bD\m~8MphJRv1!ش][yL,Mt2pؿ Rſ^Vozq6@Jʾ32çHޤ~1z?^N]xXn\i:ūmRN/33^dד/6vd Y՚&VJxkx` &Ƣ-0-<~ Y;WS5flNu^:UfO]^9%b}hv?5m_>t 0=kl~7w/69=% W8ai^OΘ zD=;??;S$-_oz՘9+,S*%10ڨX'ּH_s*R+׌./r> >ws|<^^8z*O=kdS{j.r<1ZKy5l.ʨ.?>,՛j~M[oӣ8>9;|o/{ _^~p?9"1ل_d@`G׋>rGygVzV;&ϵC6繗 +68nrk7|5"p$WZ 6k r]zTZfcƋQ\ߢVYƙ2qBv\Wr}4q OV.'JQ^|)S:*jf`tl ~wu]oiE[n6+XĬ:"]w~GNYʷMr4VkJ AHB2LdmRYf,Ѻ h9M7?=92lNp6>7g]Mۥ\rhCSlYųr H[4/Tun:20P<=fϝ'F&Fw ̤c>;e1j52Ovqduv ENhq) U}#u卸WyK+PivJWbb JݠxJܰZ7o= z鳋&6 [Y5HC!¹}\)6aGyd8N' !hj6|r6F_8ߞ}~n~O{K);q;q5N]y+\Nu6c$'hm kQ޵4uCb_RJQ_ԑ6zjv"u-ʺt]d@n%0H2iuQ9877X,w\eJ{بdM[A-dk!m݇]eULT>RJ;DŽY2Eݥ୰奕m`iؕ!wps0s_\DzT_SnypSN{-%ۚj=Avk!6K"} Uǫ>!U5=D6EU'g|vel2mM5 ѿF$o %Hap,I#7"dC.H2JO@P+c˸LZKc뜵! c>o@;97,DN>~pPdÓڈಛk_C{;ˣ5!hћ e<#[)4)^uk떞&ecNRVT7#\1}!4w2C|n<9׏KqW0ٲ).6" yW| f)sfxfy:M |go|0IEnUSr7 M'&yo5(_/ e(AܧK/WsowPzE {lE*7}/m_{'Hda~ AtO_ h[ VV)KsAjj~|y$y={\Z0Da|4]*64w:C]k?n (}SIM)m0Z{ `QWlq%q2uhMă#=S[ez`[cBvlO8}>96Df/B%K`,˯M˗&Q"/7ʫ}Aa>ek)?)nRW@=;;ԳU#t7&lK6~Cd桱^6A ߴ N'wѶ@[/kܲ*wve ҠpQio-7:yMIr:J㣱ŷ;%UdSR&x:p]2F m,pV9E)Wǡ`\P糯N]qm#ްcnڡMp\]ŋ4]+Ί6*=+K$ )G40.kmZJn,e4O4xkpukU~kj+) ăՒؽ%K-Zp7]dYg %5XM2ګFX+Yhwm4 7ȧ}`w>f>&8z($$M$xNXhY,(7Oa5X}jVcxDVqLIRQrlDP.l[L4^~_ B2!Wf:0-Q}v^Drx:tljEp>}_h}Z~ɨ]XgmF$uF_lURȁ1_?f握cՖG8, 2!btdV)&!!Hh9c&Kw]B1UE$jt @療2Cp\)"\}&l*H 5gGG*f:wggŧђD!5c˝T(鯤 ML5mb։" "SB/"%'RVHMZT)X Vj.xz١xh/ScTKSm^''/w{ˑj7|(yAryoKMeWa:[M{?25Mi6VgPxᅐ @tI@sxRל+S7uҥfMȐ+bQ%]E#أ!$E|(Y!cUkv}ؑVQck͏}5ig (d"zr’;3MZ#$8cFrZI\&D.z&AAíDPVzKej՜yϐ:^K:$[gY^t" zqЋf<lY=F+O*d"R:.Ax#Ġ>Ϻl<Ћ͎Cчn?}*l}޷'Uϻ5r5Ĝ n~|&GɌcսo'cSOEk^N4~U.YcmJ(fx"sZ5e#jLup*$٠)H@*B(BN933Qd`2df=B|feg|.<tأēˁ;-M@ٲKYd̃(Nd6cΐ,@*רk|5+nKT%gu~qxq]r"Ռ p;D@Xr[LU9;|rOYkZ'nw 8r/pDo`3=TfyK`P%O25]0 c 6 ckTIOs9O~dH6kF3{G]-+PՖ׋d:)O"e7^D;wW㦓_Jo.af= .aѾ_i :BM =X.y]c 7ToVګ8r+q!cJ>#?y?"rǮ][$\D>hlr̚`p6ܔ4_;v79s#%1J:A3u{’?_H~/W-hNRZUWA]nt*1uv)T8U(%%{_.h c"3XK`Z *hNAte6iѢߚtagZ_Xđym!DG-@b.KoNV:V-Whg )˘ʂqI%)Z.D B(O긜.9(#N2v@2vI4 oOw|ݻQ؇ma#6~rF D<7Dέ =6X}(hQ=GlL E0ZO%#GaKC(ÈQѴ֦<E3-iOgɹ ((MVb"ZK^M OJxuC*ňJ"ycɂNV&@J!%HQGB՚r ʃ 4dIfR1WDVAdw/ҁ`2cˁ`cx4M6Ie1u^bL1jm!Y|d'  r$yUߵ[m_&} O:1n}!Н` sdsJpQ0tߗY^ʥetP*G]{ޘ v:Q M-r[kBФ"X 'RTk?kGjҭ8p;^+U?NطO=+><<U-ML^[cѪL̵26ntdNc}.C8`mK?oxSj:-O*34s*WkZvi~:}~/ɛz҃9JrqГ)P#X+Gnt5 6g[T{Yw JF)} +.Jwn~z5#%hhPor|B!hut+@e8t}^LmEIiw &/:ڄ#e| '3S/Sˣ!  Mk1ӎH34(>bwk{`S<.2A?ImaN^ĄT6Kɠ0^bL H%LCzH-J~x%QN~Fb*VqI'5Z?ad8DAZEP)YΚgO(Ǥ,%[eKFxPrb0bZ1:0ARDǝ3@)Y"@wXurL&g$4ULB xXw֖!};#$vz<=-lv6-*yR<VrJf&!.Qfό9b@-g+㤱$r 9(:gz9?KU=4uqTC9|f%+HH#m!̣ǫ}NVgv xOI?tb88ζZ*+\Iy)UrŘx%|]~4,׫Qj^lcB`ƷOJzE+ $K?Ni2O?]]yř_I ޷G_ߋq]zHJEtgDbK%F-o4~k5{9}D3ɞˇjCiǿhb\Y?{ƑB8b)!8\/6y3iB=BJNtAzfH(8F6l3ꫪzcWl2zn'ͦIp7?TLRF_ON?a@A܉4'7BNdQrhe@UEIvlcN7_פ8|>YKo&=Ez]z3?k@zx/lrcc NqG_oGPDOpG׭=t|,Du2{{]}OL]M7{uVFW^I0P7i'7ѫewY,Uʮ7wv~8-rM*$<k;@'?)GT#Hw$,)yM.BUpfqVv3k[0΃o.Wݛwy i*Vln]x헷vWCkg}F>=5sw8?M*~Vw+e*ݿc` =MRbͻ.=]b6y߰LEjJ.^n{N^> ~9i!99,XY03|"Oo/@ rl*~GTPyjS;|rb,xωeĔ.46i!f~;˫}9~o0l X՛u/e%Jձ`w.'tn&6~Kۣ*˫Og5We+v\>"q;.cb Zv$3g?M~JդYqqI)fӿg HMVʀ)NJ2[a$irW%?WPۥ:Ϊ!@9ejLףoU~{=>3|Rseu/ pʲQ ը0X&aӗRKH@tY^ήGT͟i Rv1n{v 0@lNF_V-G?/[_%ٓQMʢom/e:o: gTlCCf7öj md7FFvz{t:FdMzf2C7*(=q h]%f6<~CHhe[,XOCku3'8 -=HK>Zgm1դܩxg~~>zȷOs=N}4y4j9ߦ:\t11̍ez+g 1LyuY\. пL[ F^MiBms6" Nj)d';9hc35TB! b|9斖yOhA>}j?G[Zi-2:ʪ*w2źƳ>Q=be؂*[Z6oxA[>`2[+KΫLroS /zA;z_H7oշmZ}?zCvI Н tir*~{L!]B.4}}4n3i ^/B"9q6 ;y8u ]|.\WkY/n] wcH d.g Dsb#Ww.ct_ZM+nnxGvZW2Ph 1poWawl}߁h!tKy@XZ)a6bUB {|7fu=q*aV m|kb[i8>Մn5!R⦪}cn,.O#!hfY0SYy=:,ߖY7XRWut~>ڲ|%3r ?rdž]]#`͉Ux4fwBѕͯBq}h:xُXmm>iĜ;]}ݺѓE O%=aQK? eOz7cXI/ߡe>Z< L$cVu`ղ.5s%ΪiqUǴ8L@wdN멊iڧKt이lSOg$#s,40ff^ե(LTi88G9ٌqT Gտ/%-& W9OB9 ka^y,9r2Rj$.aҊGvaeDʉT-TZ[b/?HG'GLϏn⃿'W:#T̝ͫg}o^z|Q>Qj_T"ĶNR4r鏪U5[腻 J.NU ^MSP ]{TFAֲ.F{ BiqWtoELNσ,uCZud.V>!dU&4]Cfn:_UѫVҵ- ""Um &-,7ԏܰQ5I1LzIe7j4~[/۠.$Qݛ{(.4NCj埊ʖrTa堎wM}a~ l柏<4ܢ̅P⺇ʯ9Th~96ocdL2<:I#Ty+'>Q JfeU:.p]Zn MM/}~T-&j٠oM,wyQ79'g1W8ŲA*ZLr Mu W+/# *-#FiT{oGv$Nb  |yY1"$* tGM`S8:))<RFi50#Ԑ\J@(3xSZ"} PB$Ѝhu Ţ .e-%ϑW6eKcD6}ɨ5CaTf*YJG]™depX3 Ɂ.i-RKɴC K`5 !AԌD.4KP*A0eR11i5eG ZE'(/7D CH&*,F2J`q0^+(;%7LYJhfp! ޑ.QY= eiKht@heyF erHa.{!<")#Ce":[V *㣌NqɄ%\ HT8SUe(јɠ/ict6l+nU]uP6(VڬF$2">D{D ?(Syِ r,YLt KAi=O6"Be(!zCeRC qBd4 aҀ!$( Y QDj*@=1!3{AxA;͂GԠaTU)B B LIlqpȤ DgRv na%A2<*/*JroM@DpqeHWv"-YA4pHg/u 8 f\VٕpZI]QRi Q2{x0xKJ%+pb\ !pBq6a,gA%onYf* kE~LL6)ih\w <(-Yk].I$̘ea8b1:PT6iJhs8t`߻L;[&>U9 wm2\F`|-yU&P6X, ǀ'`<)PpR7-E>;@2+ kt KViD- x^`eH43S<VhhX(= ɋq`.Cj <o@b:᣻EVdN%D$dU؝$I0 l@^HnaE`-nn|:̽`w~2$@MvgWX:=# s@CtIVvrb<IrC8*UȻK&C#X<ˌ!Bҙ An!=;F~ R"xD(d :zD]Š+И+Zέcsڪ@08@f&$/bLgU)Q6/ܺA,VWWZjJ'p;,@ "mQ:2G.+F$ frP%JI>@2)Ceb{c"Õ,y7 s.}pXeB`z;|*i5kN?@ey[!eќi,\#4Є7T*m-=\uoe!-7-IQ]$y5$oʨkC:ԆJ`Zi̝WNtZNQxf]M.:LnY"Ak&nj!f=06:7 *vh3KQc]Uf I<51e`9Aq=BU^Z~nhG3wm  3`BrH9P.#dskh+GR(]&TP=BCpH>bz;`*Ɲ-2'b^D W}\' W^#0Ɣ?" <2axaBl;B4[1<*)U,JG53⓪``mTh\P!s6,g).XZ#eȍLWjmRc:AYJ>9.J3)% &}*=TGtՔq'k[HzG]r5WhKƛZ#@K56B ZxevaY kC0_XDbiDt^Sd1 N²\ĹO G놗@\8 WjφbU~VE 0ebEPI8 $"Ȱ?tbɱ'4\NU,Xc.t]DSq,齕TvHj?&7.u*`VBˤ>vỵ詆 oR>  =yJTbAE(0>ַgGkIN,Ry} ٬et_xX/i馹WՇW7~݊ٔzbBI ?h>ieGK+\\~lJZ͖Ž+a2F$(ד3 $ԇ )j0Jl -ԎꎹN_}-k?Ua qbHV])_ js'Z-![$"H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""Ho-%@3r r^ JI r^#"wH ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H "av@DNH+؋!P[= [@" d@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D D$@D}$/rrH ׬!P[#= ZZC$7HYֶV"H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""H ""!>MWMw[i>l' uRrw?I77`j:k@&0q]gy~5ﲾg}$Lj];nKix3GtZnHmqmsnuq>_p$Se<п U_ ~|;=?߆ߤlHJyn[+ |*dV.9'}Ϋ!o}v,.hЩ?ׅ&О,$UrVV@Wf(D@NN$:pS 1A"/}OsWAo~i/Ur6ɓ/\g/ḡG׿r_oiv]cՈߣFˮxG~˺}lKi9j̬@Rq6hS"ЖF0٢a>](meuO`}S}~ݭa `ަ@w dh̘#lӭxM=^ 7 _с:-fgoowauz]ɻpq;Y]ܚ۽iTBJL|,魖Ɓ^Nߕ><]MzH ?|C%z Z}5n?sj\-)Ud  5Z] 6c[YdQu2@:jyӨ\tq:8㳗ػ(~ 9Iy:*\сR~\M= 6Cyl)/>/BI ݾaQX蓫ޓ[{)rӵwŷ_hThtK< ū:S>|ExET8<{bafF1GUxdO GzX)>h,QeS'>2j]~Xl't Wz+޻TM!pN-Dmnv+Vm KbF'LNÌJW&Q#t*4C3{!(haS筲> oǜBm3Y7r:#v\|.Q{qߨmFچ6)؃:YlA"!sO[B,`1ځS sejcҭd eUd"’` "Fs sFUx0Tv1DlU4>0 Pp,3h0&@ KK؈JG2vlyf=w/zJrޖU]w:\t+[B5oM,աݭ>\j죹P8LU&ajTF7l{zjVUْх1qQ2›?|DX9.>BI!{/wK;X^O_\ EA ?'A>O$}bV6G19 |&5oFPCj]wU(mqD OqGn/WH&WrORCwx=͋@Tm5^oR,${yګ۲EFU.ӵrl,dh齳Rl(ǁB 묷믧e5 gLf6 iOV8MWrWo؇~o֯o|E퐆7`'Pޝ¿ Epe3pX97o_U[5Svu|:Kom]74fxzO$4#r}A]_NSUAI^usRLwE4_i $hiRj^WJ*[_<$ gZruBp/4A+}!H8OL<*?}6Ay%3m."D~X92ױ%x,Y_dL>U(դ^Agx?W:Outr$u4otVx[:,GW<~^xq=wҌ% cTAm=liHK~0~v0,-~î)>DlB9 l#G*5 \w.bV r76S|.&c*s|f|̦pz|s{XS?g4zĔ㜭6*RK"V{\ōU󪼰.@\QKV3E/+H)yJvY%| u )WU+1CCyɦVEE)~ڻwK(tz,޴k7Fh~=zyR2 H1Zd ʽgQ(0ظ,e4C.x" X.47=Ψ~eX} ^-Zhv[ ! Cu'kͮnpQn{ ;`♀Aק MiϜsHFY,I4$R^yH$h:Q1+ 7$՞ߝߧ+^gr]lSߏ~w=EkMK)&$aF{GH:.Û~,^ܝkݤ'Y?ϟh# J/K!/AKV2EJqW|7dhY<\@˥~Bڿ Wq:y98e+[zAmx39Pd'L Ԛ̮C $|E%'j(>60ܛy<<..q_P@q z[oӾɛӟh2-~}Xj Гғ| ׫?¸HwMRa1PgPU''#l'UfŶ 绩z#Jd|_8η'yz+('f C/JA')>j\qP?^ Aēoqey-.Vz|L fcgüudIH2WtA_}[KC,^`pD!8i"Ƅof?GG[_q|ȟuYHsg1]0bџցUf͵ʒu/fX&4 Tg?F Ni.[:%C졜!W Լh[*M_maG' 7ج7^ymU ~pL؎x1RѤ)\lY~1ŝq[Κ !хdoM Fm;U"AyP!8"H!m  Dϼ%NLՎ"/#\A_yFY~q\ GmLHHqb16z"Ք&Fͼ3N^@N*NgrVFplmDo7g>Dw1<$CpyVweOڧ?MwT9WG-nYٙ1oh †RK+aUeYُ?p2p[9ˤծ4', O%"+eЄ( ZS+))+ZVD < V+$ـbDsk[ :a1DRJ`$hǢjgt{@:cPFmB 5h# 21f3Ă3zȤRp$=ۣ/.CEn5z:+q@P8pkZeL<|5hgE.Z|8{FE5q\zjW_k3<;2W͸׃ Aqp,~YT޴ol,ϯ^t(Vfca/&^/7>Cz5%ej >>U(p|:R: u)G 3k=ubq8E*Dx3+f`a.pROdջᰞe!7*aa^g:Bhac_F*1EMg9΃ٹ7ÍPm9|'[o9fqFrF7\c߀8_ sb'BʄCx58esm?>N2s_MI"3ٍph'M i Ds]0^aPGʒ[jN_/ޭWMrBlKTTf):XIJ61+Z7@3{ZqB)GvǨ22aJI(jigzWlQ]2a<9Oڏ/"`1ՕՔNG5L`QQLBf*X/,Bh"Pn.`qONrkіriF{3a)h'XH;£^jm@:dΗ&ubg+z5+w3(mP 6,LPT*MsfW2Suh +=#Q⟷AƾL۠?X" *G%N[-JdZ !S:CKG'A}x6[KxbDKFuV1r*($vJLXRt]={V=Ÿ~YkK&ء͛]+ƥZ8-Fr$8 ٳr7*Qkͨ^+vۨ:_icQ+ZZ#.}wV+cť{jF̈oksO+ Φg/օXdJ`?L|o]횡(̛w8rsq\Ҳbu%!7t޶ k_F/,obD7-׭ھ_5q{~vWTEn@Nթ9ʥJ+MKMX ڛ"TP"ڌ'b^1nԑ.KwB[a׷5, T 5 <2cx\M/]WOq5{"ib6J(!t2.z.-wL,y$ReIZfDaQ/"A6/ Ӳb/irE'iTϐ`ZKѕѰ?B:RӶK@%dt  q/Ls1;, ZF.袷[) 9\.")^Ydd.vj i!SBă\0(GjQDlf=a` {t$o8:·s啁|}_S>!w2,n_Y@S>ò:N-ŔH1_-kIRog\M_9דּ`ý&PDV#rQ:t./6*f#EH^=5߲\R"yz:W#;eċ=~Woy|rJSKK6_Ո14:'>}oMX E_$G8_p}oM'~gMnx<гIogoζwgxb&;xzDN\c(\BڃV6qɔք[%XTiT[")1W)-MM{֩n8lpl ݽ9KdYF{;l[בg}YS~\m6d?Ůk O,f:q8i'OAZ._,(. 3lJ+,T@Et2i[ֱXAH&ZSBq x %pQ$Dx{R\9&62(RnESF%28O@YaĠx*\;[DM+pgxd_>Oy_ȑC}F?o2I@LW_rR9 V˨2*MR8Lfz0H7}nve8hYZN٨NG8H[k^G-Q@L!FM$хmy2m*A&KYw63_ bBJx⁅#< \,P|7IږQ{2sQ;f;XkthI'광HOF:Xu"XWVV_DG[+Tꠉhi',q01!ż(\P킑V2S`]He.%g&v*;w~ՊT 7n09jal.)Oq|}[cÈMF7oX\GT'â(;s HaPN 9s;k׻;Q ^ǞfAhq/B4/ Sƺ'g/=Udét!]֊d:K8ogM۫5ћ7G\D3ss]Y+{AVt?a+Br&Ҙww&cF>yr»lEqTn?&s|nP?ގkǎW?+ixbIB~ucXcdc\!̳wAw7xi|*#:dS +O/Q|:2D`bXY^=j|_]_.)yPg`}vjڻ ?EGOx4/h|6%G=m %skB{k8CwݬpmF9l `q_ObI6Q˽BסV!b7 t.\l]c-Gڧ !fpԡ׌S78o6=&7I  VF<120ovt }ܴ.z "$p\ypA raDk Oh>nu}>2U83gp@9~rJ=Gz"Ar"qW+4'uI{[3T3(ooƝG#dbhlw0ͳ_ [cy{;Ej:ng z\Ǖ"\$!E" xa#ݢ?E^gH _{'iٰ U hZ+^ZLET(]iS TEW`u fb X6Hc1!멖[MtǍ 9\.{E(Rr2-2s:/ܝ-O 7i'lmgY5CU3}jrt;*I4 9(P)ŊH Vq1o HI]xh\2PY6ra!-XPAj͘[\T[xgu< ƩɮzdW$jNQs5e/,9PDVif+n䁘N_/,Uqy)u@Ӌ rƏߢ?;j'&<ź4+1e3j!:r&ʷDnXӇ.bTbbmS/Āc{wGgkl,}0ɳ M=BTjݮScCܩ/p $1zCJ|`+ip*A /:u6F%rn.x뜡BD D9e5ܚ+3UuΖkUPIu,dž׿:OزRe45kd8C<'Քk/"4z)٘1hr)p$L*"Ie_L͸D-L& /bټPa!RϮ۷wsw,dP9G-ʼ폇wuPC T^_3er\W$o569hnl1ʛN 20Aɉ^UDab%O,H9ѡ`1ˢfTS9K+}PD+萌c) cPTqK6*ugf,Ub.uluAu}Ʒ_bt9n䧱'rCe{^+:=6[Kw*t K謖E)A"ZZ#ì7'Ӑɞ HrceɕJ8%f&܎IL(J>&56uf@1&O)UkZ[ZEkq hgHNr2k-. e_YBB U6%YJ%s!Ȣ=uM+=*\1gC& !DmKGNf}XQt5ݣcW kDjV#T-yKL 1gΔHS8`t4:U2/PB2*JIJS`)UTT#fBs[ LʼY#~Q5]\tYLJvՋ^V/zqmFA|`:O8r`62 CА,M2qEj!bq,p4Pa%TMeGwqȭ*>#kpk_QKl_B^,Om?VZWΉO7!z:*G+ᣩ,%HJbنߴ7G~#Y.|eRѰDvD "B@6ҹ`D)(hϵon>wta TKf;BƟȭQsQCFDQDg+A" Ep6s"͌)A]l # 98S2\#jGgN&cug"G]jJSTX|ӆ/ 5:!˜&񖖻w%'h'+WzʲWL FjmCZf2ceK14 j>g TH "8:S" +Xiq kvynhz ot9{Ӽk6ϱX 7[ s1^c: zjrq: ߳nl 9Mپu͛Z[pzy͛ۛ]]_`>md~9m|CvsN|^S5^4gßݺmu>=ϚYfB_\3 7ׄQnV(v\2=&)xq=koF0H0.plM6~H cֲ8"HIE[([I86E8y $M . x\xf틧XsUV(ыh8ceJOA)9wTkFhgsi4qj gOg0u1 P(>8"T'F LJUQ XjUQ@=ǾCV޾oj6 `]nto[{-3 oI q5f^1ISf9 i)C%7ih 5(o+LwR&R|*F>T#L1!(KG(52AD&]2Qy*N\H>tCc}!P5b4RJnUc▱GO8i1Z=8Ŝ9s˭#:F 6ᆶVRU%oc ITP()} 1MLryå O=&jg>NLj9+j$)~-DS*H{vs_ߝlRp5ې FifhK8|(aK߽im5<. i쟛 *ކTXS9,Lz»>~'Y0vs^2@Ze'|=G|M'Ab((+ѧ;3L'zfPbwfaF9‚ݢ<",N$5qM,ՒdŮ?׷&ƤO Ft r=.$eV/.S?WCᯤ߽vKn\aIuUO~/\YuOْw&ax^M6LA&*K2h{kn?j4χl:iwS L_ͫ=jR:Dy2RaLnlo7MEmc$Np7eM: .$3˹4g1ţPL@Ŕؾq捜+\aIlACȲR9c6(X.8`DZf{^=#`YD1<psaR6jR5(ΰiP0vY4a#A2WjRz\Tknlev[և,UJCFKBIFAo_l/&ZE`NZ d+Do8p~1jB1K@f S(o:wt塴K# 1ۻg^_uMtK2X_tn_>a sQJfRL$F#D*\ <#twa;!r_shG 4b% )@6,1 M3oƓSߑFQH}Z$u a4Fk=.RsօPg`uE8r|X凌!<ҘJ6otrq2 =ְ~XY 8.㐘 mɮo`tA/AWnw_9yQj*% QI6QMId$O's[!Ts&*ST,`Z:KV 4 ǩ"v խ\9o&@כW=AIs&ʕD:GD,B ({7I]|"Dݰ?CTU,t| oHqr#9ɉSBV4yr&ix PG|o{Yp᧓VmRNZҥ'?'~%ƙd 5]df.NַWS$/AYFn(k39v%g$ uIvUNCetwfpgI'޹x_ 6= qՃ^qaOp&G[S_EI?\rX:8g |M;5E/zn qdrl#oly҇M@i)b%q@FYΐs2XCQ = 9T ,B-W?xYܟA tnƇhmô_ @2 A]-}eL-0Lf?&?C / >{{:Q /Ӫ'ͮ_ RM$O~~!OJO"~ykgu] j<[{[gNY3qSŴf授nEن.)4*4pge8|R>b[Ae{ܨn-ϴqcjᖇTkX4,ZEC(B3 1b&U$՚I. ;*G46oBM:p$g(hR xc@8Պi 8M)ޤ%wc~>iՂ tBx>ؐ5&:{5ظFk+QEe('6tn/yZuZ_Ļv,Zż~m!9= >g\QKrQ9xsᇡv+ؕKspsYs,XUh%! 0)fa)pmD{K'vS869$LuM$ mc}J!,E nUmm8-5-b Ut| +cK ўJ`*m$8͞ ?d!,>8i @"C!]4c{ùE<1E}wYlar~eGQV Bp_|F_/e/޿H#y"{qT `Z3ҁW Jxx3\㾃yb4kl{n.಩Z8eaco}H/z;X\[/\\e!JNG LB pCB2`i'̂={ϲࡋl~>\aF3"Eؑ;ˊp_R]β NF0Z3t!1\RٿF"n̸L3BsFn7Pea[.ɟ(e~ºy?3ΎhutBE/ [b;&!/ @7Ejs:#reり.r w*Xr"p>p`wU 4cq4'Rr*jY ,\/Yf[ KVeJn.` G#/;3? >O^k4wY?eVN'%$p_SؿRJ1zUR]STcaˏ( z6>p=w,e0z14 eH2"t*KDt'fSͤVZAykF:!S&aJ;f2RʃHzqy2pLqa,ǔh'԰`.-QJ1Vr9#Јkq1.| BF^&VӱXAFp* XEde )$(jci"nxUFR]?LAI%V1)ZLg aȱ3HhF x`yv0ǵ!pn46q[~x3$>!QಶcH,*AKD Z8sfA*7ʹ\+$G;bkۮvĆSDpg,7:D ɍWߴTlxK(XNѢjcNM;pw-x8zl Aq Q*=dopE­Rii !)yThRE FyRTF<AvU5uEd"s*r4>*4̒Bj%BCm ˎpcxP 쨓Η^b`爑q.\V|JB!1D08v_+_*iWZ;,,j;,[Bf(Pt'R(׫$lQP1sKS =Csݽ֬ޗW5l颐`T:RYяŒ^R=."2ʖ=Gߎ9^v-d^d1p_d.B_ʴ\}W& fFQjSuf-s{foJ ,MD_F"2&e Sy[o^zN^a_27^ϛe |),#UAljpOǍz^ԯo]q6riL5>0V \>wxr5nIM@Z5n0t{ T^&n'a)2! -Y[Ǥe m}=L/-bn2\b*-nmt/#| I,$k_ȹ(9?  ȵQ*g er A~ARLzDHy]4?ЍIus,9(8F "iUKs g瀭1T)b+t(ũ ^ 1[,} *Ugܾ~фu|Uʈ ά_oY5ASԭ(ivK0}p#&stMdҶ" _>U}aƮvQl'OvV0 /FA(fpbL!w #z!UxdU)xOʈhA #( "y$RD;ŝe _L'>>acGl)OQ[ہm}cOߧ VYJCƂIy&c82M2rV0ɤӾIk5' z8Ӛ?"X|`sȏ;1k5> 4b% )06,1 M3tx&ٙ]O=y>Dg`E:N[.& 1qFkœh)9B Ց/$n(ZV0rq=xV凌F8iC%sRD`otru e `f쭾3߿hdBtM;gvC{P0/uēׅW7bއg'www99,\O-ia/v 2h;)q$UHHL"wU|NKGzaU4|ˡ^AY3n]n$'9/oO`уטpSR`t$SO&e֯FVԚY&93" 4ݺQNiUG'n4"7UDs9\rzKjEԒ x_ȴI&U.Eś1|8Ot抳a_TjY]_΁ͪO? wyd~3" M68sYl8+# JJnB],E1[{[Xg7tIX+Ԇw:Evw3tԬۛG=C,N'YJ^ytrh]qnv6~,-ZL,`^%FhVd8'>)Nn Y4%4+TO-o۴Ƀ[9R0ZP}[hVIڪb*z^]:M֪U[tج[[ew4\6@VGmrr׉ԫAsnjOK~Ipv[Eߡ:rN׺2_D!>Gu#L}wHu\c1g;^\>qDezLڡظ2{F.?kyq.c`ޕ]Au{N#qKePoguy"Îj}*;ݮ/ '!X_\3ӣPkqSh GAXxƞqa,p ߍ7q1ҤiJI;u-,T}~Xt'8NXC+6Tg-Efy:8kE,1Ƙ8;N^@O"./%jgqdT}`q}48,ݚgcmTffOHi50 i\*= >g\QKrr<32@="=f8Cԁ^"$aP kQhR5!-8͠P`rHRIpxaLc1>"'^q<C NnG\ Wt|  V ǠK !ma#|KHoՃlFϳ#;YSeyp "A!]vpuh+އò~rUQl& TadZq~:g̸?.N97ÍPNg0|'-;{ iypboTHC%؉0!|xOyGE)pnGtH/8~]XN4BTtSꍂ'<5:sIĉ.&)Zfq*1d-UY~ͷLVV)b1q^\!B ys_){. Ӧ7߿{i.+0:+.g$u_ ;\lR?JSR#"Ge<܎q?Sj=LpifXi<'3 Z:O[:O|40snr4b=ƄWN. }:q7hWzN ^xse0z14 efeDi-"":D3i E 4Ҭ^w8{Ć7R-%e4,5k" A liaZqu4vJ-h|ICB޸ж&p9 åh2Xbi/CElH%ቢ&=j~헔}i9)fR ՆDa YBw $ڠc!{{#svWԳ>( Tn@[im~#6 dwD~hNCW 6X1bcgmI 9c" <. axyEG}V|߫4Փ'zգm[Ԯ]k1MZl{HHSVuJtX̦wL.H1p2DE i\<c% ~}K啤{+%tϳ'\:!4:Q"-!ZIĘ Lyţ"-yhre1`?EĖPƽuH ÌМt&6&_ hg nZzգ˷o)!4 i?D,x gO(> `* bzHxU99>ZMCܓTZEڊLZrr!i3եxZfr՝o.gg0"ù0[}PT4`ӗgf*x??H ]5 5MSŶiwuoӮ[`>Rn9Z7c6Ylݟ0g>jV`!m$SIogN HFL^˪*:c[c8Gz(u̖T>1Sb{ }q#a׻mkYd0WXT`#)Eyrg?L&I|10!^"QEg I -DhbL?#Kc>yvhbL$ }p 9ՠi0Iq ߆fjwMVzfb<(z+1Jvqun:{|?ZM8sػ6r$4N E${ 0=Lfn >mmdɣ8߯zY^iGδb},%-"D*ĚZ-Z 8Ł^`pb=uD=b"g Bg%pzE(Rr^D,Cb(4 D!28bmqT(foVHu >1}bE`<}.d؏+`k,ºZԇڞONeJYRS[҈ ր(P)De \ZuxK!DJUm CerYEŅH$4,.it0&BĖNQ;5xx4MgVK;[pL-CYjM<| 9Q|3 IAxE-C00X@S$I %KPRu}xyvIw~-yrh@7di^ j$V.=:ŸOrv p 1U$Lt6Sά|li_f<,.3ێ5jmHNadl[X;g$L T )pNkӌ:_J*Dф\*RRƹIR:H`LT%:O9 XbW,wˎNswWnX/VF.&J;_M4[<)rVxj5%vֱ$)O0y||Ǜ4t+w#+WcCD" P,&$ M 59(1[Ӻw~U:$J̭8̳ sZ|S6饌+Tv|ֆ iΏy\FO1| \ft1L-gV"TTRw]O[,ɡKz87\@#trD˜j}+E1%+Fx6&m}QT1"R21Mә+8A :gDB0(Oڄ[q3UΞT9P?<}܍U.KL{@@L렝?@9x+ON̯L~'̨B HG4$i.DZmUjjwV?E3Fe '9o,5kz- e_YBB U.IYdYgHh# x"G:36.TTCf<(֝vωcKM-"Y"nMHTVyKL 1ӹ30JQE,Fʦ֔ #Rpjqy-{DLbʔN\;-⧓ڡ]lr^g1)]4bgM56zNy97kf#}iHVkdŻbIǡr%߁ [F[_WsLiFVky#~CpottW)s{>KgHVYfaAJ kYz!:Gkᣩ-%NJbw6]?k:Q& NHDž:МH#\0i"x;荗>[\j<xc|?xl#2]$G *(A"-JMmD*SvSDTXm #t>[8H)iꈣ1b٭ kk}Ѩ|t>d6zjv[*fi~î!D=SOń ka .&]ZF+z2"%˜A*|fJ{@31@СsRaWTZQenKߣ L=Ǧ:˗&Yp2K"3H&Wۯ/geaƴc@?a28kWSZypn.b{EjRhg2 %vյ/f+3|r8F'Ż7AxZ1:g*;zծA^FKO ''lP(TYJl#W>uSAy]\WsvEy%\YSRv$ 1ۼ+zg} |>[$W*]-?%y#fktkCF;㊡rC֙C[kbe` x`*+d9fY&=6F \NNc۸XZ8Q>P\NP}yk!燲g73yrOƥMsl8b0aMl9w=)lx-As\|UA2d4ӨL!@ IyGxKCI61cIrHL:. 6~Iŵ+uj72"CL:ð3@SaE{R@R꨸Bޮ>}>{3vyw—~keB!ҡC2kH<f XBF{&%Adn?ou~_K>DP;h" %\1r62wS .i%K"U{+FL f|6-߇XI*T/Z LZrn4Ԧ)#Z4-\(+Csf}|44ӯG׷\O{Km\b]8+`2GU%zT#!.(r߳MOLԳg"Z:j=!aVN"qCw EwNHAQ8LCEbv~,c/3,>4DO>=QʹQ{~RuCm-6ZqmGKW𥉉1JW9X?uLnxq9F/:sj?i?>-7˹Q6<;\L  B6ؒ[: ܌eayp(MnfL/ZmI8בܰDhF ϛs\ j4=Jp͎J%RBmZY8}\ _~ѫ/_{~7>zeŏ/޽~I06vAݺ ܪܢך?EӼT޼i>M˒M>Cln h~ χe8Ŝ=C,WKA; U vt7Uo҄^К>MmlW0VMIKJk@Iϟ'+H϶8AOw>6v}ʏ6~HaʨO$S\z\.oܮr -2?cfי  $p'G)xvGn!x&P rbٻFn,W N}>~nL{lq7&*w=J͢ NLkB<= 6KsQq 3&H%I`# XuezA(Smbiݛ+ t(~nbc)WkqYUFq>o~ G%mY{{*S<:y\=|]y:67R֞2C;BڦyW (IDs$B`j3TI2 Y$^IHIJ)SڢoK{h39-Cs3Rɺq>v@ۼ{CГYYYbc\M1P)YH:Wb43:9 נiCq܎$o|8:*s%̫I)?b5+ѣAD t E3D\ KP${ bl э @d򜂠F貚Bzgy 2ok|Lpx1)ٻ3k, 'TClo&_/ܞ[OI:{rNkӔHكREL"kLBJT-Yp1a=36p*)c'2$9>bp:/URahzŊs@m ;r +ݽLQ ?;}y V+O/7i}%V')W/?~w%$Y$i,52HLF|>G9E81{v3,b(uP2`&JK6dMV)!דpFԃA!!wݸo{}J0ґxG Z>&(a7Q;䬣'I7yQ/jKbƨUf jF) M\yd&| ĭNT 0Κ~߸Ⱦ6 C1U[q49>7͹ݘ?{9u}PE>+]~Y}b< %&$S%hE<{|nAQe9o1&FoAYz PR6&m}Qbb*Ek '8N :gЄC0* )601,VBQ>TnގivӑU0XyRPP]u+bWYN'S^DInmG-)0Q$ ɯBI(2 N)}Z8'j,h)gw;:~b8®ioy\V lP%Zo19O]vf9v |0.ߌgɒ|/X(L9I&e>71DF'D|$ؑ`$m96 EQdZs8۱KC }gB2N[$YB @42+ad,Uaaq,X(z,<).K6xljo>qEv?שio&apq_%;O*tP!-QI4ib)a =,qh,le4&K$h;MDnd1atӈR9n4 k\P8eaԖ=j v/O {(RI K-,hMǀؗ|g!\TybfH8hϢ {I#G s0 ް.$0LC-Xqa-6Xφǖ"*G$TSJ+PgR)DӁ`L xRD8,F$2 ΈIR) keD%MbVAˆX8 yo"t\:bX\4qQybNy3ij#8wTK&ٛ"ZjR5ha:/\VHU$فd8m&꧟6wyñ5FlW_a ~OE7{aLd~~KYW߲fZj` MaN:e?3:E^̮NӝF;K`hǩN~c:D.G_hsWOFX,0"MPRrsz\i#D,8$Cޡ:TpH2K31Σ$`:Fesx # KRbrqJDCwW ۯ֣|αX~?~IY&G &  UU6HcliMPRAih۟%ghd^Ӑ^vVx^ЙN}MRWw-+mYXoݔ ưD 4"1hM"7{e+sU1A+ Uhgy{K!nBTRZ L7KhK[}}Z[yf<_fd=_g]dUȽE]ҡDh9;=b_2S R#H]?FMCs2]ٯenkBJ)&ݳ׼Or^%>yFAuy-otϣձޡ1.g~ wemպ{`h.H?5.gmyX) z+Z4Ĩ|}7fҼ>^d f")V`/حOl3Vy)[sQGe6k]Ķ3X)X˱+E ޤ$М(UTL ԓ`m 1D 2L b 9]6~4hac.V-} i#Ywa)F5IDbHG.TwvvnLc5_d+M=, &}ZawtZp4^x4U+PaӿWaTM1+_E~2j-Nj/a"⢺x/>F_-Nƣ ?ќܶ]BG~A"0ܕx V/~Ez%|KfӱAP/iRKWG4ScRR <|xO?nQxV]#V|7CU;i jG@WӻpYY,@naИ4FPZ;1.(U=+m:'9[&ߟek3ԧf|Էna>U5!\6\$kQdt^oavO{F*7Deډ ecr9Ɨ"Pgj{ɍ_!meEHdŢ"meeɑ<_Ւmm2ekaƶfUŧ ,"L>s$M|][$g>& hm!Ə5K?qS8~"1ǟ[nzs)Yp4#.-=KDd/57Z7dvjW9o:Ai!1GzP.{i&b"U߾Ѡ_b|FnFWqR(GY?VaKs7zDmN^~$fa{LF O{*X%FG~FvUǠ@[]?2^Jm'L`yb'kaT-pthuyÏx$V.rh> 2JkYb&GiWx)Lyr(K6LR; 4kJ )u2-'Ӷpf k^֝αloпqMe벨F(|^SsVDϹV+w (.rcٶ_Z$>XŶVDV4ȉ.D2Dm "8aK e/B-*qroj)甎J{:&hR##㒀 t  !d$:k4|б#ұ9dm ]{7ލ>k>4l$FVl$.i߅JDQF6`fy@ֿn #@b4 d:B B]hBlKC6: Chbhcx<I<-=Y@YrnBJ(EE+19Z)rN^:#wJGm6K FegD\ x*' ΃WUC.x&l)jSK@QC{{ 9G3C`usysFۅt8@$I o]bt=:}+N'Gyfr`>w־eɨlȠC0*L9sDGN7Zzd^9w.Cv:9%o5>LEFYYpt}+,״&`jMi[n.}SOή=6GCi~wc3^1)b}NX9 ׿&;\qަ_tn͝ujǵH߲cmq ?zd k?J_9.h*W0H5J1݄ 헬b0cݢk9fɆRjRPeOim;RK [΂Nr5ˠpN<H8Lt8,CJo1x9:pI@prZwǯ6 ƙFV/Lv{RIUpI:9^_3ZIke&/uJ֑*dz05Ҹ,hШmvEGzH)۠ab"v.H. `+eM_޹8DyS1^8&Cp(qtL6^d_<^^{7Ohra(BS4`a֖) K>@z<;+l1vf[ a%1df  dfd޳ slrƒʁr:3_=3 ϗ'3;ދ4-bԱ8Mxg4_fWwٰ >CFhI-ltS,Yg?X҈#?^a>ԛmZb&}ob;ͯ7wyz`-37}3a,5)ܾM&٭ɾk6ߞM/ͺ<,;w]t]6wywz1kV݄{;{mK8%MF0l̶r_C?aq{ N:4zYkz[X6fHŨ{R[RKA]ty?羬ȯvI`7i?fI=(4S ;wlYt|8p?20 2܅ݻ+q=r>ءvwほ] Nhw(onݔktE|k=O;]5d٩fNԈl }#7ȭVR5}x1;>Dq%QiA߀JX(l"z)E"0#r_^Pm [ox)h"S&-iB"}u1$D A2O.t p[3B)QW—YBWM6!ɻ9\$g+څGUJxsԒދ' 'k6;6'V2DQ%ihS _aZx2~HAXV:^aMS/ lJ>5L[Yhr6VE\/' 潨C?ǥbYEqz6;*EߍvwjZۋp}n1oZ-*>2*rhITjOKxOGKü Ln('&D|q: _qJKyG,5Y$(ɰV&:tBޜd2Z=/b.>L*] BM< H&^Wt^'̷^q6I 0هDv&p>g9sj.w4r=2e}f24 i<`nwc5af3UN5Q6 h ymv\[\c+?({9#1 m֧/qg-2+f8|7 Ӧs<2ǤJ((9i1f,}U 5haSF,辢y7].vpK)qRM4ٸ-yt6Dޏn%mi.?[獒_Bo#pIIdHFz?=\r!t0;!?m>w.glBqJ i1 K%7m }]xsH4(2Q 8n5H6{aZ`j8^ik0q^v2؊oTonx}LF.*e,tVL1t"Z%!ҳ,IY:|!۴Q=oyԍ7 ۉXwؠov7w7'}JHϢ oj2 3H#ka&<ǫ/ؗjy5 VqLSR"+SIZ \9 V_0ZB; QzHȲ֡P*44s.fdmUvhWzŻݣ?tj^Ë` uۀ rI{6jhX r$hQpgy䥆}6kahvl04dV_B:g\§=ˆBDC&I#,=& V5 mJdc.oAq0B+ eFlS TT6 ugG&G!LN[טm9ҒTBFp&RL2Ge\ڨaj8ˢB0 p BsI=e}!LhΕ[u-kcbҲ&ː+LF$ a}FFjպ[~UhZ/"ʈhD1&*,8QgLQYg;3֒ʵN$NB:LUD3VHnVl,hgT2p+S!Q `[seD֝yMu \\A)YMJE_݀.>&dHo:)Zߕ]!2) HB;bP%}9ȺlVfpcjq,x@%wFn~12Mx! n~|Gyُ6e45Vy(eDHQ:>FgZx8_عd~?tdcH8ಆO1Er9zXy!)j(8%4{j :7[\G@ȗBvvh>ͬ<,?2)Zf∄ )&XACQEY+ D2""&ZH0<1鎑ugl9Wrxip q9' nXΗb<6.UbVeH]BF(dR)K]UuPHa5 ՑTt[u__o3ٮ+|ryi&)*z܆rR8`CpDL{P-ALkʕ5V 8\ HC2! a*#z]*ueާv)L^f7t]w~Ӝȵc)\&I@hsO82[_')ZUnWu4}8ݮ{=wQyz~kkn.;:pG<7t\_JS o Mټ~t}w+ pmɥ,ͥd3&yYu1ùKzѶ{I¹a̓+jInpx,[[òl bHK9ݫCԁ^"$aH8äĚ!c+T qoao:5KtR6Qm0RS-yA)Sab Y2w u[S a|F*CJt:HT*s6V&oN¡άYf"K6̱5yW &!- Y7ίY>Yyo"47Y1O/ #74_>WY1 a(NwF~-륾^S^:QuCs-|.kv&.,xuł|U}텙m7kՏ[sV.û>"=x+<%!bb\8#5!iu30MKB,K!<Է^vX4ofdxN(*q?+>i`oRc֌\HKx q1^ R3u5ujL.5_Dry6 c[;Ar <iGG{5 >TH#(OaBBhxEE)paG`\sq:5?A#reʐ낊.rPw*Xp8# Һ}Lb*M3u1렕9?[}^I~S_iƞl%%ٹrPvwPJzuv6(W`& .( _bpfwYq~uYrMbl>+pQ{}=Lr}\^ v m +7@G0<JO}Zؒ'xa6F)rTeR Ya%9 ~B)Lv8,K44m =ޣ~v % CzƃnͺS^ׁCP5+#N,DDxh& N#S:u+)0kZFe)A_b)Z1uFΖ[ZΡlo%PpIƤ$6%D(%s+9Ühĵt6m¢_P0 km59UlLl鰊a,%VB UKQ(jci谨YJS̓1^Jh'6 F)(ZLg aq:]N :j5 x;#^!k/wpn(lC[ip6RL\qĢޱhhV$7ǜYʍr. ɥ~3Hh-GФhK#Xnu NߴTlL)% kCLp-xXzLoJL#8R (J V 1wڍ[C`R0>z4հVE TrW^)ǝ6;#r V[ 2371"7L#r`H3, .ā VR)xZXzaه>g+q:|%v e%q94Hp_' .s02?E;Lvd}nr-}%P_sO~1'9YWRϐYWx j.W?UqRy_ |7,dugs7R 騨۬+A@U]^ gaoaM> ty~͆e0ן"9Rw1W2:?YnNKY̙C:<2+spv($-kCgkg .ϳQ)5;d2ևu鲑} ӳ[. "RK!P3u|sW`bF LzTQ3W򼦰HoWC_06dqZD}vsF%KX%:ʲ߆A.u|w3M}57}oF--Ѿ*JV989!;lkn4-w*1ܞ\}Mzzi>W%TBf߹MNԮX/%6~3makJk/\Yd土] \rmP&`N bC@lG;m?yGYr΂e1Ǽшc8`Ĺ0X)ZbdT`M83ܭ>`k k;ĬW逰 ߜ y9^sXe9 +I#)Hl>!|7S/R?N.vVb\޿BnU^r"J̘T̤1r&R@9G+C`dikh5bbx)af+Ϝpk;]Σq>`= DC*ˌ1ۗzksܞDg E:0 ]Ldc:ք9-.RsօPޑ]e`l)^wSϬ9|J r2ndPɜ0Sn}JAHFA8؃/{L! })wEn? m"Akn`7.۞ȎRJ< p~p.qSe BT) wak2ٓt{0WPq^?\rNy{i&پz8doPM㋬i)+?Fq<8w9ޖYZp$$(efU0 CsQVyȦ*ܪn2e.fֽvO.谹&36Q[=c{Tao: m]kK*mUۍg)(n,V֤~Ha0 7)nzzhnގxRl|"՞O&etuVnHCۀt.I {S\D]]C75)Ȟ%8/Y>3~RY@oD1#?EE&):O=rlUxV}s]o_ŞiJf` y{&6=Ct7T1ۺ{~ӆ&u Vu=To VǷ(f? 턷1E&udZΛ.;ݺY DDsfL5/<ҥ "‘`g\Km0\gajE4)w4qR{LG"1ѶY[/@v^T$pb"HW\ 8k)`Bz&^a DeY^˫ɡD7vI:OMJGx:To 2U=Ryjxt\VG46Aﲛj~-W" m34~MMK0INj.PX!4w˅9 r<]\0 {SNEe0?8`-(T)UuJpw/-k-ŬȪ5b3S:Š#;me8 ds7"8vb{hߕ癇)a2;I.an,Ȅϻ?VC _zXXE`T3A(aCM%e /\(oi897 TW?#>72#!D0%^**c8e°OŹYXcV"0`1QGNR,`%Rh1yaܮ?êmS xI~>x%L+bKje&P a`Z kTѝFMFSSKD+0jMCٍ/.Fm/ȹP!#G8$$cN- )$<`fBQkI`- n$cؕ/83Lc$D`f;alu$Fo1s((5 HAL3Q[^{ M%a7F$Ru&;2b<A $^K-G-vapԖ4dխ[{a9ĵ ]B[q&LcUhT%5EʸјUZT|AYk6n`.17Bbu2=E8g\gĕ2ܢX,bTL:zb-)2us2^gqr,ʈ(boxVB'mVw"Ws_\p 401󭖅F5;lSGLǀ0 JP8t`Vi޶` S] Nz-  Re0jW@yA3 v:XmAT`A#zS Ҩh@2ӒBY`rQ0`4u'Jm( AJt+</tlh3~t[`u)@Lɉ쑫we%,6M&"^I!Ir'!ӧ.}tT ;>Im]$J:}u6F86=.-ESnF,y.e}t7@Q,R"ԋI e@7-y3qbȂDy'p;(@-"uQ82Egu]`mRiIAqF.FC>2T w1`Dhr`zK,Eae +t7&&=G (:K/R N%TB`>)'][`-4fMjAr>䏫1r~țk*3 Zb% -颅 I`$eo6:U ]m6؁Qhj]58Er8ZV   Byq0gwC{ofO5p7LJxKtM:9 -5HM'%1%S: # T|D(zBW+z۱Ѓk*ֵҨ ӫXW*eO+,Კ$r i<2apaBO(1<*f*5z}AZ`pҏ!!S6, J ѳdVgl,0s#SH>XMPϜe<(0)H#RHB1X>=FrkP"\ufffeL7FP8!P56J Fx:X@a} 3,+@-t1E zABl|~)Q#*`Hϛrۃ9mlH!DE˥1ˎ";FfQf%Ф~0"tŒ.QlAHҨ5깲5>];gX}Kz2X@5a/^/BeQN.ō''gW97iE䠸CT>L5L\6_ɵ1zn:,*D6=^_?]U#t71{&JLg"5왈h{{&"J➉6338V=V'1ܠ$KCLc}_X>gWQ*wo|_oGA+vӻ] yޟg緻sm>|ܚU^R2ꬍ\z`:c5 $ N2-wF-^l(oΎ[m8\{ۯɜV %OQGwzmD2q[1n'r;0R)'_dIb(2$mյ %Ὂԗمw̹~C$8&m^}uǞLGZS[ލuH)Y|}bm ,%un sL8sz?J/?-g/՚vH f/@tI7b}rOHQL^-l.I>N{!쐄ֻ"i~vg'q~fťޞ `(]8d[llva\%fTŢ}v_Z,g\ϗ\֟K]/p^'W7ÓL u09g\2Z\xioy"xpOgXey 2hV}M[[bp՛p' ȜFۃEyq6sf3:=Գ0[ヅw} (adJ?껖ZmIxvN>K޲sC|zp Q5]|{z}*u?>~^]gQ&q(G󷇫rl,!?Ab3GyGę`gRzN#OOs5Y^c"8kFpznֳ vY9ukubxe5!92:=!}|rJKԯ2O\>crPyub鯯{→/{};x_^ [Y&4̃CA>; N-[[Z:˩u˫]Z^yL'yGB;Mo'of?|H/~Z;#.6]sN`O"U[կV[ ͗38՟Ggn4 q? h{ ~cʑ/*C}2]e~ ヰχ`w]V2DĜ.Z:/`$&D%UcII2V C`DfQp1Ӻ_h}d]q)eD {[^P򺉤yx@KrZB`S gJOfPD+Ofo9Ӟ$⚼񭞾;8]˳Q>zFm i|ay yAn--m٪Rkkb.6q0_qΏ?bATy1;.*xʣxo6wo{nx82طdb^7殺F#_fB]^55GWʁh}ӆ<,EptmT򞺦XޭuFI+ՍQoizVՏ^?¼|φɋ{[ԧ06H)㠂7 UaJ"8ׁ&淞/j]euu'ʺ=:_GG t]mi=ncmlM66ƦtncmlM66ƦtncmlM66ƦtncmlM66ƦtncmlM66ƦtncmlM66۵1Rll~>66WgcccW Jͷhcz|zXS߱#g@a퀉 4b9DVKw&ӓ7dzR|׭ozۦ~X'|v|8.tXY!dD]".Tz1DnnUКo[] Y5LV U`}ߞV_+\ϵ#NN] `˽p'#{gY[w !bPTo][xוּiW* K-'=c%&J~t#)םuzWwzWwzWwzWwzWwzWwzWwzWwzWwzWwzWwzWwzWwzWwzWwzWwzWwzWwzWwzWwzWwzWwzWunKAE,Vy d1؏ߕ ьӆy)ŀ Ѯ} QY]%K!f#i\b& Fڡߐt17U< ¹[!:9@+Pѹߟ}=^쀯ݭk;4w.z/&mÝ3g4`$Md kE^t:Re+Xyib,gV+ǧ,e=/?{󫫳oQgOpQ} bi<|<O4+سDC?yOgZ{ɕ_)w.rA {W籿>Ŗdɲ-ִ<1#iY,S:4HjM4qAZJ@~n|Oԝ|1S-oaqABs74zҾr:;O_èuF~4`*_"u<"`]ѰYv۫7;=!ܻ#~|–v ag_Jq09.ry%0xH9&=Bgu@) Up[JvOrs zOL fx{m2y0hɫ2yU&U*WeL^ɫ2yU&U*WeL^ɫ2yU&U*WeL^ɫ2yU&U*WeL^ɫ2yU&U*WeL^ɫ2yU&U*s< }pjV"\HA[ruz13Kx1sM;3(`62ydزa@H66$LNBbksU}=U N׷.'PsROKiy>StD"G*8.U;K7C7O<%;ʐۯ\RpOϪ OѬUG-6="ހ+.KNZԎ}MoG%F_%1Q230#|CAol:Κyc:`pEˁa.?j5TN4V][q`أ߹Z鷑ϋi0~YEՐzG*wI᜻FXɵ<6τ.U/M& WmyOySb&ܴTFxb'i1?DYkuON"r1ts;dm/Fɜ@+ql\֍7]h&9t>5Z3+UL;/"gz`^KA˳)^ɥfV֙z,SBE^cVqay":i;.X`VxydL1Qa] ;n`C0;:v\Vds1UlQ T/te aEaa43L]-?cys/:E;.y]0^Zz} 'pjk<_r5Õ=~MU4~27-;[gC=*G F1~8!\^ ْ`;0 5f;9:E{G\/rՆ[@-څ+N lk1|z JcFmĆ 8j̿k̿1fHz 2z뢉.zq"¿9\'=AJ+cvנqo$;χ=G$~dГ>؇{޹ 7ץ~:i XksCsB>3!oȲT^6 7T&5kٌF.FzIX\Ҟ-`LzŭW[xgu<giGfŮzdw"+*NRb&%2Wbڿd{{Hn]!71=|b9}) 6 C'kK/Q^ߢ-)=%%8 rH)C2!" RpCjC ;I.ܜ{nAOJPJaDZfFǍ+@YГ Dh 2"Bb Ake~ïzï%˜A'jvI|8THBMpX+a~],tRT21Mo3THY"!H@8Yíit*:/֝c1l~<܅hظN!㤖ƌY;4X>,~jrgpҡ L쁦pM|k;_Ν۵Y]so3tnVOn>̩|~?y|F9 m .ڽ}Y˫_侽ϯvvj]l~OG : >{ӹq۔)ϗ:ZB\~5q<ʭQX2$E N<y4[d.I^%Rﶛ#VuHN>9|܂o+Nko+l={F܍MV@ZjC#,q5$ SYKX Y cCD"IMBaIn$E#5Z#_a?ceU~\M/'NCH;w-iNɝ=)پ V󝠏䂊WNaڻo.=m< _JH.8 d#gV"4TTRA$[6ƒR\k $LPX)y(!\1³&D-I[_Jgcp=݇TO_\NZK$%Hktj|D,eTF<%lh; rT HG4$iSypHq^ɔ /b9PEK?oAFQUFhsK/DW>|(n 2@$HZ|C,yjMdGJ F!VpYJC !9GUhJDq,tJB *n)&tR=cJ1_X2N//|?{xljˎoS4nn8qmv^$o-Eީ|B[.-9E4Skd&=< | ـ$9Yv\W) &1a0*]4Tw{l75]:NڪVkWh ^pƈLQ$,czl#QC(  OՖEp6s"SveqQ4I.P[i~^mn}JjtVɵ=)gik7R뙸/w]H_fҊT!k}\m/B_F7F- ,̃3h:z  an>_<|ǎg9&4`] Tߍͷ8!wY\+""ab+Ai|SemlAwqݍuwP>AIkj 1=.BJFk±|v!4|mC 5avS:`>n? },ۯD}\/:F5CtTz lKh⊥$irTHN OBZMu|G+8Bgv< 8Q/x"* 8R58&F`2\.9kz)3nBu=;]O)XDL2Hޢk~7~JUZyzRF;Vٻ6dUdd-&"lOp(~8% %ڦmXL_UWUpx,K @A? 'h?YѸh||S@?>-BOef3 ({>ya'{ ]§Td2gOfVQ= P^ L_ He7o*oJ*֫mU2ߖUk X&@OqLNu:@Zhj^oq|A]-!_iN. .,=mi|qh8aok]S:jhsy|׏R:8v_ăXC.6TBiPupXb#1$q|':zixarq- :[_;­U_g̞ 4$iô#M~9At{)}< V r"'^q<CCBnG\ Tt| +c%hOR0GicU=e5#:@G⻬d3ztY>_6?iSr]"aGO #ײd~3~ﲹ-յ.s/Q5&-yy;V,- R|+Fce9Vqcw~ X ߷ٯʰ9?d%;b l.ZIL,O >>UּNCKyi} }V^sVlaFFcV| 'xeEg2jЅ*z1@hbEԩ09ctZ&Oo0tF;Ar xyݜ&~h'<^>膕 ^ [½s;=Sz4^~ "ˉQh.rw*XpAjpˬK}lxr(\Hʈ/tʜl.K 1\,*OTW%86sm([ZH"}v:(̗@K ɣ?zkR47o?dEԪbRܳ\WKkM)F5%a5ipSu3>:/?v NӛFJjr 's Z:O[]ޟJbvdDe YOEޒPZ1Ke}Ri;Sp/-bm_ߞiZ (Uu'G^ׁCP6+#Nk! I]@,m)A"x3E}b1o{ Hn9\.K#"ݢD;@lOEz4X'!J* )c!/)δѡic*Mwp{-X(t˪4H16ƃTz;1(Z!2$5 1w{O9Ja1*0@!|RiB6K1Pɍ^ywۨD 9wem5MoPn ! Ɯ0ʁ"8$z7|ZIPic1&em xgcxPw OvIK/1DsH8T/+Cx 8$@uRT wMNODx@7J3; tnk& BZ_:3U¸\qXt'x LBbҹVpbA2%i+,\(~Rҝoj]X(Vcߣ\i6}WE@R`Wg>OLj9+PڰTvFoEVw_LG*0.kv1𥔷!.0&P.m'PB^-~}Mm꫹pt9lUn~+n9rRu%hJY KIp[ORކ^ro!3w3:E%=҂vܭA=n 5+r,s2JrO&GVUR9c6([ARLv!o2{oamS)"bFl#΅JdA#+"r wkpC!fHjd~sVk/嘋Jx->c~9k\,$)W¼|CrNrG.6N1g݁[Qtq)E<ԦpK#XJI=0Dϭ.R[i-qD_uugJyQv͚ o?/ĄVLu *jƂn Hx$ǘD5vw@۝t{L-ӲPnVLjE0c:f(E40dN{Cⱎ&őSzxL.G,І B+Kx03,*͸J@F4d +?Vp";W̺7q V$(RJSlb4jbTJo0XF#a0,7y":c2EZgS `d^sÉ ѓcn^~{&ng˹~Z\z3 V枒dfK0F\|bym`ThABM.TOIR2<*(bGْ22!SDʄXRP[~T{ 8YH L -j;!|֛.ߒ˹WjOkUvM*-,8y{ix.OUZ9Le }KgUdf]u⇋i<8N!0K|uf>q?9ƖR_/&+S3 B-- Z[b|uKgm͐fx{3=0Q)` W@6gM.[*A7t־*Bfyd i4v"܀"{8UX*&QUV3??ǯ^UǿW1Q/`LKB$A5_4ihi*NӼ^N.} >vK{k@e^|aUdO _W8NwMЊ t%q=oUwߚ&'XJEXRFL^KECtt#^1nԑ&gysʃC~v4<2PV;GJQ`쭐_O=c4+KA!^"QEg I -DhKf=}yFɶ z 9a>8 ꭥyN5,UjȂl8H;jBlMl[h=|{u!:=MwWz\5=t`;jcoN0}WL]+Z61|Ѕ[cgmI 'N)xLf-2< פ vj:Rjsfȵ1OӍr  i/>CڡZ7C˲ہ% _lwH뉐ctyJWrn)&R(:˙g47\?ɂqgD,}bpXĆ3jֈ(RPQk1(#cC(w{ҩ!cFY c*h%{gqoR0#4'(xAc'FFJ.$.gXw-~>*Vi);ÿ 'ݢo^.ъ`L^A> ^,E i0,VIw,dS"֎N_xV^>{ O^Eã_h| Z6KmczrN 2(PZ,lLK]x)?w̞T5 ]uh) [šuE^U<Y`+eC-f:Vo~[oimVcL.â@4RrQ'g^e}Vݰ =`5fWS?pg4bBn[\%tˮJSZ)>&=I܁[ɽFBUG}mE y29/֏Bg3߶S b2kRP-R2ELkC?38sn-5ZVh]ݰ.5ܦ/t)u.~oي\SU'XL2cŮzeEc_ Qy,>΅~JiJCuLf'ifyYMWoz:TTPJK ǞX,f"M7_ε7JeO &~!q60]v#u&"gҖ{z.(HmIvs9﩮+2iE] {׻3-O>V:D5v?Y\j{7$zE(Ѵг zmVLv$6F%6{5rFEKtLEk@nSΗrT/Viե+SM/A#s M׮j|!\Epo&giMmN&51HqsZ[KCU!:Jbt1M*4FߥUS)|VnZtm1j\W+H)_Rw+Tm %e76iRLj9 pT=|.Ʈfhnf(:#M1Eظ$ѵpGK8F}xouZhcVB2G†힪d6JÛTw̹ca>MBȥa1N{'l~%B!}tBbu0b頿 vyqJ{"T2+8cgDɘu!gk?CHI$ͪ"hj*uoD $Y1ڢDwAGo`]oaxXkQJNiB0Ǒ֑~B_+(-i;|NYbѥ`D)Ȃ>4JtᚠZdO]gE¥:itAڳRȮB&Qw_KͲ#oGDf#_f| E USk\4sk84iZȠ x(u;kᗢH+fD$' GcB*&6/ I;L')Q!&!&Dfϰur;M>o:@GVZLAG=$]I4ZTUFB>L: !'icGef5%t_8gP4D.(2ӪAU &deZPO5zOC"Aӱ*APP#H.Ȭ!}åϪd!T? y E*(8m&k)uY+N";>mtlhxt>rm*kD^ =`Ye>ZbM=A!%xB>h AJ._>1&!FuuMZ5(}pYf}JB;뒄 Xu+P(Ы*!U3ڄ`1btJ߳yp!( g@ @Ae#kvpqAkOl: EBiv%&HTT@PqkUCYeE(3A6 ?Ftbl,'J sAn1V =liVYI$e(mJMKKުh{942YIXmW6 ӻ5 YT&ZJm0&ܽWv5.^mcZ:Թb$Y⓫FP]!lJ''44z4:J$4Nd4QGmCYk Q^$zhPAI;)g =9=6"|Ye'EYF A2 5@JAO!Ƞ rw=ZT}f=*-EcW,BvV$⤩MkurSpMr^ag] +U@ &ezBE5Ff,(*Fb1;L ;?A!ZsO+IUE :ƀ6m(VjҢƚ5A&EJ=3i^& &ci@Zv!; Zt:-] v%S=xZScUѫ=8E6V@P8U8P#,]QZ4j䋞P!V&Eυ4CS"p#fD9YJD9kk1 8%mCɵ+F([1x @pQi8m!X7ϦrшUq4D2 J.ȎYxY 5iU.$#4Pb鹓:k|)t\/U =]wrt %=i,^U__ ̮%Ii#cR+}ȣT'7QˍwP;)zXP!s"ΏBl?[]A : )c"bT<~Ra$'`eL}$Qқ$@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@_- dUGD ԣ!\ Dj@dL}$mL1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 $@L1 @V*e <$ZGC#rœ@ƋaH / ٙbI &bI &bI &bI &bI &bI &bI &bI &bI &bI &bI &b%E`{L$G4<zœ@޺/a+"Z`I &bI &bI &bI &bI &bI &bI &bI &bI &bI &bI &bIzkkRܹޫ'_VՔZ=no776Կ]ZEY_.?A5 pu<#&j^bM>ZdSVx_m;f>EZo #oޥ2=[AesVb UEPyXcD)f} L_\`1  32wp<׫Bbbyq9~n1tBsDo@<h /1A.=ί8=G@뇚z5gM=1৶HIM>" HAn4ʐ?<Lp*_M VQ Cѣs~aGcbϕ7Kϴiw._^}H|]Z MD95l[YrFij0Z!zkB%i At^ylcP:*UcZn=/ .>鈀^ LݜܞO?=OTUu[l5dRҏw~Pzj2KM_?7= }L?I+1 a2ِSqX\pU`n", ] <5on9f{ݘ:s·땝e=,ːCx[Ы ؖ-1UQ^^ӷIǗN߈B~y5똈~^D *[PΖou~t Y|Z~+]cjlz^.7O;x&g>dnnWg;}=x7x_. q}YyrRGBȡS-[%\xj,(1UItژSиNFqz0~~ӭwz ]lc*ʅRo;?Z_/.'}M; ?B s| Ke@jEo7zb1J7}Oܞ>{A?1,o1]!ndw9T{ƮvWb)H<*cO]hP s{c{ga܇ѡc476{~IU~zOQ Ahe%Q&Pi5Z1wQHK-O3fHn;2E}ѩiJ1%׈\֫j( C6TbǮ%.*7)[:@tO;9^?>oJ{ۏz_||588. /GM@#&po~fCS\tW\rø"qZ J??|yVn}3G2f.nI(JHNS6W`p̯'VnUi:+0br>+wkCµ~YUxM*# rRm,({ҢtgIi%8Ub,5A&v8ѝWۗPo&wX2^͞XWj祭<{glJ@p_6c\h! 9-jR\#F\P Tnsk',yYM ݆ڥؿ5ڸ,o&|~7ɻgspo/iTG8:ż*@vq$'Ҭuq.f+hKho?& uս ,?I2 lNPKPlӆkO{]-z"HOVUQD8e਍4UբeDYɮ]k덞Zl[DHK2.pvUsrNT:S Rπa$*JOJ-5'Y:{L-u\( Whٔ2}Q?GN7,/aL^ eTD3́iKr~/Uc.r iRc+Qz -R 1 JqTQ( Ȣ`brX+B"Ify.W8 |I9Ե*ӝkqmchQ,=fsɲ୼oNב3-dR#"᧠o10A(с+ˏv* $D8c%錧 4Nhaa"{&CDlb aK)b΂wDðqVFXG 2qvOwmGs׏*#UJh= 8/ gyBH(1BTDSAH+2.+0!P<U-zg՘IDk1ؔ]AfΞL[YA>@/wpLDN"锍5Q)c8@Q^Y#}ܪh$ N8X^v`Hj%N@.EDO> ̦ar2 LYH5tviTxmoFI:<6M\\ M{\GYƣuuͬ$ֵ׷]׾.v jF ϟMgN1\B2%" GeT"uсx0k0Gt2}h-JQB)́jyH%bdF/cL&3fnΘd(P^3B“7 7X|'kk黣V_+7y' J zǯ)[ke rpQ'qFebKV!GaJn4F(lC!ZBIbTH'K !#a:0- 9r[#rMg7Ƕ_1υkgC̵E˵[( nrC !+Ejo !]Ip[ /r!g3 x @zpG=c4safn~X 3梏C9e[⣈ iPlj3HGJt&1s8 Nsyk*P#QzHL|z ",iP΁$MF´RҀܵAsMgGNWm/Vp3ufZmjyxNEljL1pY2'3.-_<_<~:΅XG,lj;r6YsE=Z'~8nRYS⚣A^R2ZiRGⒹK.U(Hu$ul3+4'YxAoYr -e,(=R AjnN3MqDj+CuUak<m~<jxJ-K @ UdD2ʐ@ $H$3cKl9TMr^BYIp}j pGME`]Yju?]HveHOjU LX*QuNUqh6l$kId¾EY D$E^\(}|7i~ߌWwB16b {kzpCU9س|$NIٕCD]^?+~t>кSrz{jG6۵'ݾuв3oݶz^4yb0ޣ祖a4lotBkzzv^aT+sUcF{⬛IsY/6 Ǯ\mec[}8ԺVD@Aٵˆry Q!,8K¥a'TԒ*'ӥoX:/]s8p.͊ݧ|gz䃔_ 1DJ%BB3LJr:FbXJ% SۈHO&ˡa9$LHMLc1>"'^q<_٦:D% # DÀ+:?7RAhORƪM|O]f]-uQkW:E5t)+.:I(o_UUQ`8j8WF?Z? ?(^V6Cwk}]|?mW\H3(.>> _+fA.]x ԿZ//q|]fRѺM .kذ>tŏ0JBиj'(Wީ9I#V.z)6gj{9_i)Y_F+&)wnt3+]ht 16aYSn1anzSUVeT|l Ws}vFgRqlO*w4f@dтS:U6N{OUŊZ7+t$@ džL{%Ty`y)h0՟RZ4 < _Iq:9HE*'t21Ő k"hklWئ- :"qg!MڵE yED@H!8!Ҽ:Dh)Q)IU[dg 71 ,8/6榚_tv_ůWۄ2IPRfC!"d)IeS“dB{zr=11 ֊L^6ZJL2ID:XƜƠRbO{Iګ =Qjm~_\\B`3-^[?%zb4;OB"O@OZ jJehr`O7g<8 iu=ymG[b΁D+#!bʡu ĥ|1LxX8{cț`>0rq=xVӄ2Bp-@R}.]ȢT\P3߿hdބ-6-8d0tWfvwwl6W#ԅ\tPXznbp7QZ"&:֙O iTH/ +L-d|V*~8d|4,NG xr].+p$JYSISl3 SK/_lB7m°xEZ<ۢZgmA٣!zv0R)sQ"yhH+5'^)lj0ߺ겵\qc0WyW6 vQ[n7g{tWT,i/)r6y]V_n+{C[I {sںyW[%tdZo,XT0sN)%-v=w:~n6m4O߰FJazΛ$ .;Sn";qj깼E78[ k WN4ѴtVhHo)FkyMxRROB8$ 4B^FуLM>JFYo~E & oA܋kPw71p;.HxT҅+rU*c" ƎBK@ ިz;2/yCXռ%TE]q"Ee.i'^ʃ\G5fIGҸW"ru6nW!;Ud6nPiuvjS -EMu4g : 9!T /XuZMz"%uahi2Rq`U 0JlB̤TkƤWШ'бwV-x-zv</R[k_Mr31Ki23Rhnmd:O(eA,eҐd1LrGLrg)^/ I܊)+cC<_&NydRe޸6͜/E*4h88DžB:*&C!Xn4)W!55qޓ#hT JS|ڞ<ђBC򏉝:-V&/%6C9su !2fQ =m=]dMJEϩj3F 1j (t&!4DK.q)iN[$)]N/g*Os14WUn_X:$e>`L>p k=|&i2|goߖ_'}dv!ʔ"/gP*Jߢ}"MT8Jz! 2C9\&bԊ>GG'NƯXuXM;ŦvgOݦO(&I!7D"32kB9X%(%%)Rm55'k#,1RTZviQtXk޴3b֢157o4ڪE=/jp$O;`5AKّ|@\$)O=y OjSdb4,N,p$5 ŢR%}$"rd^`,rAw7^B `k660{AӁO=y& #a< 0.0 nz.Q8,>Y'T]@1Wx[P;|{q_rc)%b!|ѮnKq=oVs~fU>N(6ePLG>+beP%oAY]Ux @^YsMuj|Wm*f[˓wY[?yur2lPbF.Ѫp wEd~tX592)j_[._w^R\Xd$VKQzchku^/vg30X&TYQKK=aE֌bIm(x]$1I! س twG|4P;G9r[xZ@dLA;MRROj&,1i6 \)C)h5D2|lPu'd>"ҽ4cȭtszFN 2AfEu0 FJ:0"aڃ& 6$\DVP^xP E^F*zc 6ԖիgD ) ʉ&O8ybsK)6eլo{L%Z[YmoVCcWgO;}"MHm &.P6$ǐy(L&tْD@fѫghr{$l'דT&fyt)bҙ5htTߞa|ao#P_zjO^ps]*q~]qG`*l K7qBx̿qm- spwM1X֚▼!#u - ]E%݆B\$=MMk dv m&iu)`uݾc$ޡ{zmݳփV)8gq +qǹIgy1cҦ*-+B E X<{d cVb2oy޺~Zؾǡmgr㐩EP<8fsLA4δ1DSAD (^=E`HII6k"s3>2TFJǥP8A=bocKEG/62mm~8m#.#2f !,.qe#˪,\}Kx,q,DO֙j>7tΑ;b5+j7䑿P ^o6-^s6ӣwTiZP{/ϯzHDl=2$[;`,liH4#ըSJ:Y,R&P2%!ȔUG2Aq(4ic0j 5. d`4T%3e CǓ4S+HeƜ1DoWIeMI\oRŖ%GS]HMwBXs[Bܕ'޹dڊj鬩}XCq (ikGLKTh\@I3¬!dEU[#2e 33S: 1"l [|A/6"=p &zRcPP%"#MZ;"-#\} e6$%[ƪe.|VzbLBjDNъhZfvj*K"4eLN^Ng.%Bm`2]ӽu*^"(MhQmϯշU iΪmYTe.;?X!0 !|*tZhv|g>"ձ8 7_hw[pԭ,IK7*nQT$U7F^(bPneB=@'!yyIi.;Hznlb+qqޱ _DK iڄ 6wsHv[=o[Ltw+f{ӥH6ѿ ck:zqvN6Hgt*lfG-ߞ%iGw>J=VGvt[1xטk=\6e-O~ג<4W_u_6桾γ8^$y YOPnlu|@qBy@O5 kS,#qOq <8{+ĩKw%Yw2D _aL['NV`s_$&j1=76: ,0#LQY.jg0K !1%DĤ VKa4|Oҷ޾d\~ ϼSkRfG Zj\v# *ΪwiyZ 7,rX: t"XSM3=U!ݜW$ OnR\A }|=Zۯ\Qӷ\7w69#0ކ&Pu-ZauUhȎ[<2RWs+18b'*5M'Dmm3aɖ/,ןW-ki۷{YɊvh>; _uvճ7u'Y(7%\ƅB1 wHyh˿jFDj.)ý}֋GB)][ /9)v8/DpVfŰ s#Bt策ĵJ#uX&b%FG8]{t7hMn,`խ tt{}BGGy; Ao=oǦ0}]::7GVmĖZ9nzM I.hNdT6`K:s`Mͭ4J @֯,>\w3YS£ ;tPIDA"eMugX. %߹dJ"B1 _d,:d58DÃP( Jr:IYT@Aue iҢ: +#'J1]0sܔa)2heStYE0[ӀU;0RplLH#3Ҝ#r#zyKGhEIxKC;"'{0l+l]( _w.vreiXH|Dt,$6iׁu-k鉕; ka=8D1, vl(5Q0n\>sRql]@Ԅ"Tv4B=6t@öhԃx̔q1B'.U B(i&&rҢd.AB#B9LR2Hs4h"zƈ1`޺r z )`I0+'-ӌ~@DZ*y&sx7",DTXcy c9 1+FN* cB΄l5sEAq!@dɂzeXMnKԪky֏o$@wZU-!0B9_r]v~:i*CTZʀ֨9HeAS9RoN s-W֙"4̽I8HnlD[7>];/I;,}:v̫&׷o`!7e0>KZdD&DhVR: eu8ʘȦs%H ؙȱY}i}a6u,s}HSz[/S5Y V7in/ 1ᤪnr'Gx650L:0G4bG@O wY::Zs3h=6k}Z6ttGb|CV\ l`vo3^ꑓ]l? [>|Ve0P+u>=@e61pOFrzQ3$dȨULLy=ǯzCZhu+gsCd,rxsډLtBhp*>}䥁,8^;rJ&Ųۑ&0J̪+~54{ڠS]yM p1 Hx rt_1ĤQzn R"1:SXёﺊ& q#qž[}40,pڅp) FcD5! cX.{ۈvLF}ha>CC:Eof7;f'{K}8b˓7A%q7N|H@^K&``}Ar>)y"LgT bp̃r7~ޞӁ"a p~`͏UM/qd^OOG<]Ys]Og[eA+h 5Xu&qHfΪ_$4'Wǒz<2˲ӏnռC{4J5`k?{Ƒl /;Rl$:@kdHʯo̐")4$)[8GuSAP]lZ&ڵI$lvU~{jokͪ,6yv]? STjn59jjv]V|t_Ӓ^\j W,#ꏛxyh:aojk]2E%V?BGUQ+n~dv XdC#(.JJ!N1#/"imcLwGiUrU^}"g) ̻>pjs9DOw\# xPkvT4LPr-/g|[Ni%&udZ%oMl-;ݺY ԉ.<s禞əf=|g^ =Xj7ƭVHDӔ=09ҧG1^^PGf\縙pu 6~U@" &tlE,1Ƙ8^'J,|iݎ$3؅$.ǹ=s;ąG<.1uzP*r3eJkP0<B/Vh7=)g`z !G$ 1Lh7p䃙w?)}O_wb{?{kdMhsH)D01:RLݹNÍ|>@ ޜ._@Y`A-6f!BCw0!}wYSV^&) Y&şybz7 KO76"g fe W.f7V̂( ?ܵz ^9oФҲ ,6 Yq GH\ARr2^xr3·_'?|w;f%@18_]WsK.9@ԆJ7fAf5I&75556Q`ZGDOٿ\Ż= 8VEt{ a4q)K |{nKr@dUcaq wo|.?^~%&?\; ̂KU ¿@{tޫ-^Mz5ojXW.G]Ox. ۧz([}8[ D韽_n?}7U }hySƵ/NM t%u==?^TZMaW aWƗ*D tlmU4]<;DrR0HmgQ+uBR1F$`fSG(7c{EUA#Vp`Z*TRW$Et ǑGv| 5QŇu&VDuja^FrwNWP`z+ H˃U}s5)xWQkƹ`1X1]^r-" Ғ)Ei 5+^F2볇E=M=փU XcqUX qptPѕR_R=HU.O `_^YkED;IQA)3ZX'}JGZ/x2(mˠk tɉ6#gBZL`}RխE6LgR(6+g0B?ɂ1֚j2cmK_1'\:!ڣTs%`A1(w{ө!cFY c*h%{gqoR0#4'(xAc7uMR/Ў\Ƴv->ùYb{j/Sg g s\ .Wī%^"*@ꯣ }Z,3d1_C5 |qa^onp-\ބѰ0`8PąGxp{[溸~҅~0o F 2 P0 ᙿ%vm]|}TBIKW*_ī9xsZ?U 3 z/^Wy["h<ڹl8*vx2O(_;m;gDz{Z0MLa5E&'rjrޞW%>+yeXN*+ł9 ˭.":F jUVR1}wԬeN]w&jJTG3K*GsSZq7.ͲZ?Zo36[ /Ocĕ[ ~| o*G3SRm4N2Jdy6N#nBȼGmHHS ,'Y^̛:{EC"auRy tUl뽖-{-A$a eHq4$|XY ߸~} ./7kS_~ rgZ2ѥ8!ݗ{~.cV@Ϫ^Oߟ:k3쳿y嘺''H;eALBp*-2"#4ݝK]%ɵU^ıt4I{O!ESYq ]W~}]^]hӳ:{Ǚz$$=gLI3Iz&I$$=gLI3IzH$=gLI3Iz&I$$=gLI3Iz&I$$=gLlyr˼-"VTgLBI3Iz$$=gH zI3$$=LI3IzFZ@ZJ23Iz&I$$=gLI3Iz&I?R5[w GUQQ,h9>}.uv([L Рp!.Gx`ye#L$] JQT3ǰN:! a bf^/MNˆ^FcD2Y+냉K G `h#pϘtǠ ]ө/.v3gX }{QNmn *b'tķ"%1 ]l?l8xRRx }КHu x f!αf-4*k:\@QAsgR6Utt;;x#zx2bSⳓ_nG"[dY޲lSeoKMlIȒ .u|d3y8aoeQ.60":rFBB,ZXJL!0"4|CQvBOY k:R0`V2PXroHxHɹ8kFh Kp3"'c k/%%\2N,s$au֝ᇆL"qe8ðG |+|<5v_; Eo(s`)$h bRK"Z9yvykBz4ef_~䑜 :2%\A+Wmoɾ0q0fOϒL 3y KnӢH\"3XRZe:[5X5V|2Ƅ,Si"`x Bht@urd Fktw7^BDŽnqHdJsvb }}rѯ%ڧA+O } (;TS0:}͠p~01- 0{b` J0_`% !3WPѱp|<UJh= 8/ gyBH(1BTDSAH+2S`(C*x$ )0t[X1c2b=6)5 c`YwZ&0=*An>}>Ѐ}x9-S6D% q=FsJǃ@88ft I Xq8I$ S{+%RYwߨjRxmh&:ޗ3.d.;O{7\{ݺ&ꜛUIRk% 29+C?{t0E_IR+?~1B2%" GeT"]%a.`d`2 tۢS(GkkmHeg~?6X$@pHCA[ F?%z)RPTg(Jw(qwgfHv zDp;Jp Tߞ7u{Jo/=B9ƒ¥F7w..+e6d[wהyemAEީB`c[>Y[7(Yw5$2sI6A DFcU)@"AW#RgI\{zmճVFb4 3Fc /99x4c_9Bb o%$Ά%dȐ dсF_٣BD8Ѐv}M&5>g8xM#>,ig ɂ#6EBLi4J h"b∓Sڠi$U 6)MyԂg|P`B@&̈́NZyߵARgڣ_Ƹ)(/ڞ7x?<; E;sn)71B"o7%qE!FDZ=h-stԕpUϑ[_2!Wݏol~NʕT5g1>z7R RB{qqFH7:2YOK[Roh%mb0l2+ 4dV2),z?" &H G&OQPQQ]Ul5q{1'|)4zw2'DKZB5@JH ǀ3˜HeƔL~Kk|Sm ʊ_?|~Z~}8RҮm 9KM\w-w}#'_v-Pw(sk]zH@3EbLԥfrR]0+bDܪ hmSJ#Hؤ^/b[ >+`eT9I$NL)w6;p5lJw y]~ 'Tj */.alY-\_\\h`%-E ľJ%Dii#(NQ 1򪞟G3~ZytKSlbv8 \A]S_NB5;Gt ¢} nnZJ!|+7ʐnQ$VQEĿ5ݑEP ;PfsQQ=M\lb]/cc]n_DKi҄`˷m"j쨘tY;oAɍUS4y֮vl~2hnRflݩruϭE;'6#ܛ@lYbvb7u3 ^b $<rny Ǝsm;Qzd-R`ON&2>]?4onYqW-p=WqŽ:3D(ƵHjlu@qyQ} g ~0( Qp ;w8TSw\KjO]o̺c{ 2jBu lJJ#FGŤ>"g:@c@NP:.*- KC$Ec#dHўCgCL9e$2'ONW ^:mbgC ,P 21f#3 :o:BgLgf-UQ.i`t yR(P]Ǣ.?~WUEs7^]Q]Og}A z߷_+Ӣ\4IbEw'Od|ߴ+N>dMNN3BѮ_Oǣp25UrԿZ3(b]frϺ핻Ɨ]? ;􃴳+YW.(o߳8$ώa!.h1'M뒮.T׊ DU!r!_I,< tIUL>E Nk)-dݢ|ںI 43$b3t3.j/\QJeMTӵPYv/M.ꎐA6[+Qf4oF@v̚ #UY_qJ : ; eBOL記)#8Z>Y܃I[`?#ieXt!?SV^g:ybVo8ҳζs6 ZނaDŽD_{ ?ua^::1t!w |i)zC@>Z`ϵiu">\b`S1+@`"7DaAC# OMpolg<Tt"ͩS$}(7uv:ˮL 7n)ɹ!i,S*XE1KZ *FIq .d[BWF(~܄j+#H@j ja#!Ì;JAtIFr{U=01٘ ߸q*Q6(obPh,xƄCwsE&aTFE;"!GȻ7l3l](|_w.vraiXH|DIYHYׅuTYJd*GVLR^i}L$ wD~D |X@8P#'z K|D%vlTܸ`4j:}S)Ԥ-B -/'uCN&Z TRbK/-90̐rcM$?f׾ߥ󯊻3N3;ycnE/sFʁy I!8G!yu8h)Q)IU[dgI/L>Mw̑u_ͯsĭy& 7븭 m˽*^wz잓^!c&o?6tb|Cz+.J{2n~zo47 WΞ~PJEIPJy1O=@<UuƔBYz0 s J괽V_=gYÈMK)DDQx. H5ՔxsjC$IN ow=%u4qZ0|Q+Vz#09~՛:WOWz[Ό:Z&xի':4y̯cS[Qrحq@Go,劾VA'c_՜JG/XANRÇqjChYytO>'6$1f ʗ̚EpSt[RƄt∓М1l8{9DzR:&l,ƭ}G~sq m9V2EW]~'_rLZ*5!eZ 9\>AIXCbJ## cr jp ߊc4kDEPR+ PK) 1HD#꒳R*%懒RwκGz[!;$=^dq u`z=utF93o_ZAWvt}='gG:|g:%?!JubgRS}B6;B6k][o[G+&9V߫ۀ1df #IiRCRv>$u))Qy,:}.7j8hm .*;6x=))MN$wN7uܬeʾ{ެz0ٳD0RmO?}F1>ݻO7w'K V;ubtζϜP_$,֗ߝn޾=/ i=H/=T% MŢ1-NG=gK+nR=K]K+iZ4-Z*3N·g?R还(ð=W'sCE_8;/>gojO6Kv^ɫU5K|]>[ݗ)Wa^.wψ\<;Ms7mUD^JnhkeԯNj,!le䄉 Kq^sJSOb0ѻ6X- _GQjC!=C*`s1@JMRMh9uCM`̱V8tR-_i1c-X(K|Amv9rm Ƚc[.s,1Sn5/0H)/r]VdR:ZL B7,+Z]1<5qh }/k=| 4SiY^[cIX{fY "c`B$!K&p ]7*?Nd|Cf׍w*K7M$ղT+ ,J% $ad!CN]j@ʣɇUшFu[oWD;pdϔUɍ?8YfHh(JyI_Bc,# ĘP 1ZMz\kM[W<^_p.ݓg"]iOF<فPcĵ \6d~R#"BdK{cQ2 t K><>\0˨sԣCиW/1@$Ab"02y ضb )!%c\IY!&h)\1ƴ>U=U&8Is21I69ТdX|A؜HJ\Mg=N&ၹ~Wd#|WeGm!#:d0q߯t'r ~LIhy0{Xuԯ}؃_ZҧTpŁ$7BZgoX ݢzDzRJV[NGaKr}67lu~Gi.緾v~W{p{0eyβ"ch)?l]'%| ݲ NKŗ eg+qoŀz,^l6nIv ]-S!m3H\`5$ ͍4V]w=k9lۛ>4FF߱u C==KSnZd7/[|kׁݸE/.nM6MPbnb{"Y70^_qL| 6MrD,)2m"Q QƱ886\Q+o]WHu;ܼBjcE=y&~㌐7WUzf h^ҴQ>j"hTTj2xa~zWg5x*VOjX d]92DjTF$c B9.RZ+1} usfJw1!$B꼥o|zi> ?:t+u&C>3.ā(1.v_gYD\ m: /H6*Xb--c&_ilfǖdi~吤fYbB".BC8c&K:4rSV>a*.zX} ا 9–G`Gl!׽\B0 QK i@v!lDdo1<$II@HMVȔ0*+j5q6+j:(`);!P:٥"NEܽNjG֋||lMY*o>6h8`B2dkK1IGD 3ϣ3s,*EUh,jMLO 6X 1;/"hsV)n$rf2R-c5q6[zX-2v-Խ-<-?,ȸw斱ų/~48M>+YT.'w!mXA1KH &HJJ*o+i(d#\lJ(oT"dv̺,KʙJA O#jb8ޠb]lMo{ NB0p BFkg =eyZ1dbz9WWY(Ț!WĢL@&&V* {4d1H m]Ѹ8aKW1X0 >vP"Bo{Њ&K d" s93@4kH'$ 'EMU,gHnVlJbg|d P H,APE&'8[%*.vvXV:(<RaT8l楘K+0 m #Ju9Ubc;vC< XE g~V{ď/4zJ|wh>h )uR*7iӍOoVV'\Q}ff3RkL 3ǠxLkCH >;B,Vn=1GO(cP![6IvQ-q6c{#KvS6ef'y-*]V=6vf<@BH1?SBͱGGτYT͡g4L0DY@ QBmNStWZn[]#(]3 Lݢv˝&4}X5-ޠm@X8 a$ J9&䗳6d$l QzT7o|l|nJDŽw\h.̲Ȉ^]pǵjE7',##hc6 K=DK܈5 H( ,FAApe-/+(g9*Jg6i1Ĭrf"q,e"O,}> ~[֏L6T)_~KkĆG1sATB|H9ԗԭQ8cd8}f> B np\ D}9*xC #Z_Mw(mhѡMB= fvCYUg텬jˋAo,ͽr3\ k,j ^$RVBȤ)ЛeXh8 ux,|{};އ`]4z ]C 됢kHikƵm~3bes\:%nn蚣anq/b˝6ya6=EF-GCׄw 7sK)KT{#TM#{ܵ{ D.ѭٵ5jҙJ[4jhL|x6e=hrg$hg#ު"z+H^E=uSdgpO ([ʴwh6jǻӆ'5YeZ7|)yN[N޷;6~;{L{YQ$\m]ܙvՔ/j$PN*ߗ V`({@l6:p$g(hR xcZ# W)%;߮-1O9}vЁ7v$pb"pה5+⬥Htp8"Gc:Hxu% t$/tv+i:/Lp/8'ؚr_\O~'xdR#:OvfZaHq i3OB%U=ţ%?v89Q[!P݁zK{|fQUb:+ mäĚ!c+T 5!-8kP`rHRIpxawLc1>"'^q<CΦӁ0rF1]gn1x6X),y$D{*g6VT"o !`.ٌe)i5i,w)ٿ|4EQdϛ/"+Ǔi*Èr9gd~3ޏ^ilǽ{k}e{K3epΊd\̞WkA̜۶ce`c#]?,(ƕխ ?Zs:K#ЩPe8r"\K6GP_ʚk `"cY%GB'8OY)\엽4=1k:ٚ pIUg'Et'd-y?3٠ڨScr%se(2m5cx K% c⇦ٍ=;{מTZmTHeX`'„"9?6*MQ3;/ 퇡k+lH h |RE۝ V8qAjr l|9s"Xͷ)YfQ(1pX4KUUi9?V\rSjÔa= l35}BRpŝsa>_~J}ǬX}VΦ%$r_rH׆R^J5ipJ?9AėA;~p<'MP0R''~wvdM[e W tp~D#?9R, [WN /9tx`jE^_=|ZڪN뮷<u (E ʈZbEDt')f`8(Orl$K@/1@T;z|ו':Ζ9o8bK"9%eĉ5l.-QJ1Vr9#Јkq 1;  DElmmXAFp* XEde_!OLj9+ڰTtFoEVw 1XMGM8`K]zRKl .N(BtX,x>a?V \}CW_OŨs^7:=ZV)WDzr-,2ugԫKƺO準inBnqNKX"[ ˶\}zhs)C_`NF4.gXvc% Gރo;ݹ|WK5K"61Q*(H "eS,&XeR~eϡyTX-!c‭iU"bK(F[ :aFhNPƎxMg徼Kiezvޤ-YՄ^˗70T0ؔ񩓬Q8}[$pY0]1=^}&)Qey+OA33|nP˅VZK<[R?{\~QMxL23ʀ[^~$ddSdj5.w3 nL&R0ekaRBցY>FV/ʺrY.Oײ5ӣ4=Q&Lޘ`)"e7)(ݘ᭾ּ*8YH xIZ% osxUNT** KInk՞5h&eµJjɣlY<cz6=x ήOF&vf.8VkKbmE)Oi\Ԍ; W A#1~HgmÐax0!qG'u[ǣUb`"7JQl^!ĘEhGʚ@>O\:,i>ƕ>:xCh_UgQu9sS ǽ?~8~ަ?>wǯ~8D^ LK[S/@x] 5CSŶwyo3[ƽ}帥@;pd JfGjJOYVz~jVvA)լo~RW5MWWw[B1y- q? QM&clV[}lJaCJ#68R$#Y@d(Jx-U|`t;٭~Ap.Gi Ri@/O4tN`RGSJmΌ6>橠>`=rjW_|bGY=6F k t߇iuHKvƟG>.Wχfrɴ˭e'pTI@XQ[3tn09 )${Ѽ$#QMe;2ɏYBN3MqDJEcD2Y+KM-a$EV1MMgKOƗW_AritRUpWYzXr=:B[6jVE5H! 92$P'56!***/S&ƒEߧzRVSWDc@Dg1g;Ja81w6ۙxjϡ+MfNܱ)䠕6|s'p!Oޜ./u6,Ve*VrK/n]ɮDR}jƳ.T/9Sہd0a3ky2܆rR8L{t$S.+=ҽdJȤJL8AYp{ȸHJA0X x$ )R,˘Z K ߃`MFS2t]tK?{WƑ|8%#ݗnEp pe_%PR!)j(Q68꧟ꮮ2;v: e *DD"6Ě̔rp:`ny̕vo={B>RF?~F L&E[hi5 N/Оt .TW&S'A<Z2XIK*rRH43ѡ]糴X]_?YvKtr<*<@6UU@"r<݊U]VS3 J K7'\YS`Zm)]8Z1kQ+_mww22Ĕ<`uf`6://S iK#O*w5wY=돏5$-] ᧯1p[Gu J0cWmܮw%뷍Vj>4,]"l> iEA֖gmZ*Ll[('ݚQgm4!֙ %[M;G6y ެBZn/v]>|xnZxF y@l7Ap3-pS90ϛn{c)nm^r-r2Skq-::qV2{ǽRNW^-ueJ/=gT0 ܥŶ^D)K]ITRTdBO ͝P7uH=b5?zHkYɮh{Es#.M Xmк4. v&SLȹ04oqE>>pP78<ܱb ob:.JяZMR_䶡|GNLBBYd\镐Ȥ9e(=RE&"m:NnT٘'za7T[Yu`^cݥ}!2ҁ;̐=(UR=!5 z_d9Jrm<*R6:@ Pqk~` Q 5&8x}\f8ߑ} .BG ft0eVمxz)TrL ' HӢ$ŅaB-9+~IaD~MOPk i-z9v.*m!v AeϺ+xb]oK$3d]"NdWow5u~~ ljo+tQumۭJq7?p@F}sOKƓIXG_w Kk[-1*HBOAY!qd #[ofJ{~P\qWTB>i,gpj TccmFiލ&?3nhy@VOT"E'Eã+Dm8_UZu8~X ƭM@޾N.Jo mqnOYmb"fq(^ B꫄3 g`m!2$@t.ΊGUcGj+y -z恱2Rj{ݾ7[ԾOwF:ixZO4U풵`=K]ikjݝ'{p|W䫦pNK[4jێd]Ke=K)!noUa$oN=u'cyP_`Lz|bC7KfMߓ `2\nip*yNG;:mټ7_6qJD,31Mb\57ƶ7E.\&'bΦ&IT%G$Ky~v9!5i'M ,XY9ηZXЧ>{|!z.7CTPyjSQ +fJ9w pjJ3 LafAAC jr(iz^b_^xXs=p0dBpcL.l^ӎ%C = -pJoWdw%97^莺tg*yw֝e'j <Vʨ Q (NhM 欬4jmtTLZ3!2z# 4,l#<yat)1%c#dH"=S!M9e$Ғ *Y:mbN 5dKkcb7ո[Gٶh'ق_4Ի(ӗ+BQoE9ZwUUx2j2Ί%\VGE5LiRPz;s#T-ctVbnj~9qr~%Rb懓|\{'g.cQą4_~?ۯE58ꇢVӘVo]𹎃Z#9qy >؇qxoO{PJBuꀂ:褍x N7[ݫ;t-+)XhKOZ{\_PgF]9;Vۯ7ZKWɄDUDN.hcge<oVg6L@$MŬЎDDDnHerK 4cD.45Axgˤb-ђw2'c91t 7>RsVgCc-e˔ )YE1KZ *FI-.pcɦ0i͗,.sZۚ -HqR.Pg0{&%S$ O-Kevʍ谴y\ՆQV@iݟL1ƩF٠u1C9-e޵6r#"ey9Y`=H6ؗEi+#K^ɞKb%KZ-hjVYůU_ 6P@49$  utlJnC׿%<(0g={kM;ҢNfSr:(Ҿ:椃8I9rUTd :* -9>&|HՖҝ.:wC6vHnd Kh"8F:*aTHsZe6nɌ9snXZ3[4i.躲Y7L5Dx3&}~۫ǓQdO{+noRbm5W +}GhV1e7%]E}}75K\+XqTnXLjv=Š9Qu(tOB%+LJAg݃Wڴ\\r}(&NimT +IŷѩT˳&0Ӂ l3Кmv-S2{MfZHoS"+Μ4_OCOf<3}!|˴c$U|C6<0C Qŀ^6eWn^ ]{RJl뀷eĥ|l`V|Xiw.%Sarˆmnb`rђ黷%3[= \oM.[-ē\mO2ap͙HQ&8㒉))ʚ:^Dn^fd֥qՏMQ7T W:ZZ mWՏf}m}I{iͽwP`홚K>_TiVMa<+*32t>_Jp|yQ';rrGibe$#XHJom>nfuJWQiӁp2}]LjbͻnU.mguB\1;ʉFZHtOܺ4A7|0yܰc妚o5fAr?w}?s?H O4)ṼEAn ?~вiҩ]]>u3Ọ|q7ϼkv=nroH/8ƿ gZEqB}SE^sT W*.~ҴKilnq1ΔACt4_1Hhn1[S$?ْݵ9{\?B}vsʏێƾ͊z䒒029'upfLL=W^vS58he% d[4: y:PCTsġkޣ*ᾇQ/,ꡗtm(ϸzV C8hYBRcimQJEU]LGOh5!TY$1V*:'y/x!dF蓧E=Ng@:yd9|bx!Y%o1x9$;^n Ap|`bYmLU]l۷R=R^ zEF FYY% o*ґ 2¶Q/Sc}!-5:|e(kWve׃k+{4>3еH뉐fΑ刾^G4*t6W#e=^ *r*(+T [e 6K.|N`v́h{\Cv{dz9-_Ɍ?~:t]B+z [ 0c뤜mzr%Bȵ)Mヱ!8'% /mlgc6iԚ )5D!!ZYqU6JXιEX&8O3qnld5 hQάSoVe{q83g{ INo| t )m15MdYqt-RѾ3c/CJPeP)2$U2dw}&{MY#*Ʌ̄`6&d1[^E}ieJ+UShZHSxs(umS_f(G$; C}QjAtQEU+|?D*Z]FA+4Lc< Y佰<+t7<󯪻?B}UQuN9]HaɎcu&MLU _Rd|+2?Y#?И %ȭNB)!qBQ)p %%j-m{BQ>ܸ|:q爄.ۍg '1Etu=ﱏӓnsi0O;oMJH*jRVZ(50cLkgi^,=Ѥ&7o#y/pEDT8WӸRcXcF㓙9&)YQRC@&,<0-~wE|$pxX[9$BSZo+;0=ҚWNhmvΔ2힥5@*ȃC sTR!uë[KM}\<9=1$|Md9(M5foFIo8}U\'d|;XTJ޼mذ.з,͍mnYwK3gFދn54^_TeMK{O1]i"TJAdsELj,yN)Go0V񩽤Cqnۅ~aﹸ`FbqHaZk|{xs]Y?^ j_*T%nDΚgu@(`FEU`ocZ^`zc^%t{M'>i!bV93$P$|B OhrҌsȕ&cFSNThS ANFQ C|3qv7z oړ+4#$Ow`K|y`rWv۪`C?eb&$WcNB$9bDv[3+-\OVDN^Y1Tڸ_)A7|q$+zƜBEp3] oGv+ !lM]Nfc @6Q+9:_u IlIMg{_zש@/u"Nx {* 7EqO:> "۱cMYSƕbӳ*QEMqFXi…hP^#rKFzO8eMU[EeY⼥D 0OdW>h 2)ja&Vf3 2ΪB6rQrRVs3ÌmE * #*f|ShI$+UP'Beb"5lT8=q͜ua)^w=?E?>wO߽Fv|Zu'q<o{IϺڕe:`>dØwuuӿ>m<}=}99wt]G9 ٥͏3~W WN'wk>Nx8! I'鳻}S^n]8;Ԝ$1? 7'-S|ק͹03]ǧhLFtŪ[V2Vop do͖Y[ d i"s\V;,Wmɢuwl܃w1t buh/r\nհL|aneWѶ輽c:?(#S^^cߪkv$!5)Bs$v^ي@N\0|YA6E8&Ba7Z;ܵn/`fׁ-^ c@c)rkF-݀w`aXjf6K6DO2|[@؍d&@#!n8 W(ύhtWUwק'4s?7gj4QdG {0lG[pFn@;POgpvDa0Mk FΦ8rׅ \*]$t8zvOk4 2,3q,I-}R[p-Itĉ`un?~]/P\6`̲xJ`RWr^r5 S(3 S%S$([%e#$SJmcx@p]o~oCu0@WHT[H<=?։ݓn%-o[:E_Oe{Wx}74i_-ވ7x'T_LMvp7P 5gkver"fQbV$oPMr~(g3,syIú7M|g}Ͻ#D0kr;G$lsX) UZSu!tV@^2I?u%Ų_E[C?۵n8Gr_,gtpXW^:3Q0􉧫[< ^`BB)x]S\g]βkrn|.)nZtWt1MXg_3Iyb.vA-~0 nVӈ3?]j~"74= 1Ў1r(0~;+JV*KGHK8-ktYFmR[6,]>?u=sF>.I/duljw]M/ ߈kS5,DՕ`jx,jUJ1IE KPƬE (>TT.ZU 4Ջb޶B-SncxI+zXBqsJ#3,TD20N%14kt^ƚb1|RnRZt|!aۛÔ-jm.IiR o!:IY&X&5$LKc[M"ghƪf`j ЍNYZ&,B)F|7Xt>]#K2ʲxI !S6lmeLBaM:JqEۼ "xG2UVR+Cxwv6 4ĵ+D  V1h7|rt3_|eAb*[7I`ٹp9)l^qzk%`eAk\ 2E%d&o!MZ[%sB18RvM$6Gqu$~k:AjqJjivNas'餞5/f&#$3th±h359Ycվ,<7F|V$\, @(`VJ!B QȎYW_٥$Y" Dȗ gh|d rNaO-OXƮEE lxZSЮPsYe~V3 QMb`֥6#ʓa*tp9M G Ѡ!.-Cuh⍪L%e2qYs7o AF3'#)kٰhVeX@L֗]" (=U+dMQ<%,)l +\mB3URAĊl&C2Vm+v4d4F@\œ"X!I0( !aN@ rt.|lUD2Vj$=/ `WжipI ESAAS@dm{r:C9@B!dA#j:clV"˄[X9" 7l,mAq,m{25zr6 88% SM 斁3?xjiTꬋgQK-"]6=`f9xY"@$lwͦ o{Sj`[Ct( [e)Yҍ‚T W~8QN+UZJP+B,죠0>p+v` \OZE TNHJ@OsF<9. )Y_IZI+^2J_<m]nPI렫 粒v*, 0 ΘJ"+e.Tʤ,٩S/sW]QH.k(VZ7UVxozT5g'>59%qJz%qkH$g\g'myȓF4' Ҍv+}z;K钹dK$ <]#x%ut$FxÈ!eAʉ$?ښYHVDK[N3%qʦd*|9BW 'At#˸]H-aԅ7=bR;:Pj5^ meĥBRؠ3vug[@s[""Ӵ3ɭ)qKKG Eu rBt~!7u)xSKK0_-^իiQ*K\kF7G]\Ѵ)G߯a-$߭\Zs~x՝%u}mӫםX0|)L}rj6ܮv-P6zZx}W;)F$;nÚo#o,o`!46 fbɇb'{v龇!1뤔J5RNp_SBvzUwXT=*Љgu;;uO߿yM7_w?}o޼̼{|5(ܟa)p(:[%"gp{ܚuk܈cn-|s_#7w1lyvroo;?\zޝ!||z:vYM*AnKO@lXY1CRcS*C(z)Uڄ!NsuE:c/F&vऺ^n$[HuHig+T5̹CS[q~ꃰSh/Y78L<e⯍\}?o_6KA,'ij<@&,Yg֐(xP3#<{'ºǎ,,tU/?cd~LTj~neT ]ŕgWhoxQ=lș)d%I^'[P$/I89`[$2`21T;P4;}C#Fh)dzkպ/Ϝx/9Q9meGfaP_t_POQgł bmʸʾ4/3)GRj/^{T`/gς%v0YzA9@)%!xkߋZ}YbQ1 -۰9^/T=b Bc{bIOy8eY7\B@UZE ΅{?|%r\Ag ,xL58t6 \96izw`x3h@[E]ڳBlü/W2aZ-,T58{Q¬man ~R1AvY`; lg,vY`; lg,vY`; lg,vY`; lg,vY`; lg,vY`; lg嶳LK"W\r5+!WӮ=SCrF%E 4>Ep#Z`TTJ8*sUa8LσQSz ȷ>ȷޯvHޢ_{ _oȹ~"ε,\#9ȹF5rsk\#9ȹF5rsk\#9ȹF5rsk\#9ȹF5rsk\#9ȹF5rsk\#9ȹF5r\ε!N,d=Upb[,02(B%V'5xHYܑx2vIOO6_AflK>~|_S.*Srq&jֈ Wqqr O&$NL`%Pm2 N8BD gŕ'qx& Fl7wgvyQqS$wxkh%UOIdi$=n| PaP'qpWYy LDK,1gP֤,.2 J]_͂ ~ϚE"䮨{nF<K4C7/t݉A*S[y5 v/FP׼AR*} )%T : SzJc;,޸<x9V0[hCh5"gc6l3`FEH37Q MO'^r`=1Ƥ9`nY&ad ڰП.hf@<ʖذǀŠ+,| "~H3#X #aq$,ő8GHX #aq$,ő8GHX #aq$,ő8GHX #aq$,ő8GHX #aq$,ő8GHX #aq$,ő8җSNj=? -5{σjF\K';'NMy9g ZedӾ WAgJY t[qwXPg ,2M\%ܜ&pb,h߂6 ;m2>>FtHXa3"nǃJ3&؜|HbT,jmFpZ&a.\sst ^_CE\T](.'.@³fKI`Hs_<&B\9K`͂$ 5Whv-9P݁[\x:>ŜeΠ}]Ey=`is2m>`=?WjodY7x5F$Pi tm}]Cnw%T!Ui[ 0Tt$ށX`gZY9qj&D?y|^m"G)NAYkOE>Lj}9w?΅3.;\ai6Xթ..TzVqeS:F bŞ׉CVh򥰆J:IF!dQVNhmځbk5tFyC等嬝֔n2uxpC eTDޠ NsRcHkǥ5u͠WA+B:Awa8}ys]{\VKIua)'ur2 8__ͩnopZ?.^ޏ]L_[+]+i%bPURH4+u~"hqCγM?+L* Ukew~](}L'\4b-Nw' sHP$݅==UikJAHx6u)v3/]4p宧^U#a"'u&s7>_K!t|;v ؍iv~%(1,lQ7?b4t fŮT;>b]*BY#^[S:h!6)yEj ?t6?!_%i_T\]?]xkzE\[vyu#ՇrF V9H*}Ò[O ݭ_\z8qG6y[`"mh=Ҷ\86p)K!T"9^y\(a+Y1#wJڣ'hdN7-2":eѱ<;Z dJK[ 69`Sk>8~&U\Ǫ<Lvy܄w.?6adL&9ǍI+`g.Ҁqص}"/Y_,J3[l=FѣָؙQIփG<H͸f(Ar6YeH Px2& i荷91MC6-{V f#J \DNokMIim8.qRH-jH+QuέR K\c5IS8azVVf5XQ;81D$‘.%XEn4HsZ 1 \%^ bֽP:!mΎS;ָ8o.xO{6O=BG3F(m;vzOk Y>C"YY5b:i.ʯ4R;E$Fb c&FoƼBHzg QKmc+uֱAŹaw`OߞOXcHJ&Hkrfsz GtBȟ1r2*BWQviDuƂZ9 54(a)結.,'䖼[i/[Qn#BED[z^^Kz~Kr Ld$!r<&#%h+[_hE=c=cJ1_XL3//A!-:͔Z" GyFhC&{6 uq&Wu &1a0**`t݈RxV(fWv18k^[^G;qE-ST8)`eYS}qOT(W& O 5~y5!XDGօ$A3lܠŚ6~cgpl)8#BaG=.O*!Ke`5>WfB4 ;SHŚQDL E?NP_~25'!ގb PJ_;'\s !8Z Mm)15uD&@>d_a$3…('r#Qd`sx# (SPWXzݠ~1ݵjw΅n}>zl#QC( y OՖEp6s"Svemkx;asTʬ}ܾMmÒ9:j}ʵ>gik;sׂUvUKĬ$e 3c jjG?-er8Fq*[j<[S}3G L9ۉJN/0ЯWBzw"k~= 6Io# c5}7j@f]-$>_[v(U|/Vd4]]W|fR2K0e~/T&9)d? -?^>@#\Xd$J'u!ӖK*&ԟaQENܙ6MgZt<#XGEZK i7zs@^=;:{f )&srUR!M3!1^%${E.,SXGHN:ӆAԏ#V5ylSZw]<ò;')aVRM~!!f-_3h ƬaӢ{`t&Fw&gl2Eds8sz^B4KVEP*B/G3\ֆOE,Sj!f+9B~Rye2a{`Pu68 ;dKgtFq;Nĩzw6` jYrC*ݤ9FP?\z/`D*Ji',yr?rv'$h# / FZ u6_Qy]V芓9QhNN񶏃Ջx|g3qJ}HhISz)x/ijN_& @SPG 42Tsoq@S?rBI7OHBaaE5C/MYZrֿ^kj(5\XjiC_ ϐAKKpIa8_=FҠg p^}f &92^ۛu#)*N˾|ꤔv6֊2Hdm!I 8xekgo:@󻕔 16B~X|Ԍƽ_mƣ1덳QN|&+NÕfT3IQAW /$T1(oܖq0pjo/W3˿vrqnWޮ)vpnzV=hrd9RYFQ pG7Ϫ0 r{Ӎofso*'Y Qw:=ڽ)maz{>6Ss`j XWpUjan Y镽޲Z=5'ٙ6 #ڃĵqeE@m^Q@G,$ O%E(>YWmPىV;Ѩ0d 4!ۙU;MO-z PijJdnvF~ȳ~V]6j{/Ŝ-flY-M_;V=u!nԿح8K=쒠ɥ.2q]rݵyё~tAG _G%jT+)P-0Dk"jCVtx&u#MtVXKϏlPO@p,CYjhPd˜j$\л~aߚz+k}&ur|I=67(mTɓM)`4QlS>Eh9 #N1WU[_B>0]eR7~|͢mQNh]'D9۞P|W;Nx=ys1v.xg/,%Q"X ,)%UN@Jb/ZGBkJWW6vy7H+1{B3ȬUdFFfOH)JK.#J f#A0p2#iHFr?xne"eIAcɝ;-gԳY?Hʫz)wAnG mҽ:P& r@X՟8s(de_TֿEγn3^8*P`<;D/ɺ$vbkJYۈ?2MT^~9YYOE9}+[D;RضlSU\S[JU K1>29 7ŨƄ gM'*|QZNQ]pg?GkYIƿ.Ío֬dGM5TY% F;~Y.%S+p_〠2!Ĝ&}6hXoE p+G1GTvhχzWkMۋmnFMm/[l-9t| 0*w W#J jh1*ɻèjW\NyZ2*@]L!0YI)buӫHo*U"b [p44l^JNy4!0, C)8)*'¿tŚe%K0+_ [K!kO^*M%ˏyJ1L6Iq6/Q72k`0i32eEI--;--.,vv% "r#ѓ AS/19cƃ#hysY,Y>53 gkQ޳$ :Z)9biig]̬Q]?-1(!$Rd^zRǩCA&ʓdmކ<_!`$N&P><;~ s>Y,)CR$xi 2IT~=^0=*ͰI=OZo˜O OIJ)$+&D, ,{":zQqtjNvGm$'GD&d, a*j:I(<ݚ&m1[N"CvȑƷU.m#J)Gr7ɨɜpݏ`dNމ< ~|[gkuqv'VoWn0 nn"cps|JA%H s48 ޙX{΀0A% ~MI2cKj$ ېTFRYAyBpH*zc|֢xhfGyrHާ;K{?y>6^MG]Go;99o vfbL[*^i6;=ʻ~hFjkm6uq9ǐyVA'-/^5Fc3Te;(0`ɽn=.EY:-.[3֝ݚ?қ.mf UϺP IuGwx|'[,~mic6ڍӛZA!c4-5E-yCB PF9+uz#PdIPȞDr) Zz-6L܎\$!e ;5N=y,Zq=km=hn)hXs @/wHK)׎C&ԾW}.Q!CĢHtML,hx 9(:|*2CLև-_`x48qF4=kD3hA#n^3Tn2l)Hؙ6h*(H8b!~ bZ `6 <=(NH1i.p {'K:}͒CY/zcDaMY] FHe.;BX] #/K>˪qeScч͎cчzVZ"EqQުm]#*wn9]Ҥ=~4yE'Z<]OEkΦcznj#m4ZKjU@sZdk%< Cf!ґfVRJ:Y,R&$hV@f:uuч"_Viw t-V?cS%!Z`4U%3KY`̡IpZN1ga/רk|ݔ0%vm}>E} p;]BXr[LGAZsqJTk]ӊdr[h7C-}OߵjaB;[jP-V }f~=4pF_Wǫık]۸NONg뎗tL4QfR%x?.ZX[U-5OkHAa8 Lc!b/Qz4.N&)~mug5,Lj*7p;r^K#nmDzzM(Ơ耎PҔ0Z<x7k`}yEkm3ƪ7 /RqvBLBfz5E,MږZ,.iY>[^_\LJ4*e u*^ PL%Q^lzCXOZ˲(hN5 aX#jmj}՛#%> 7_R A'[ߨ E%VzvߗQ,`IAj'l_^8*g $$|=Tf\?fz׵_wzxBV#Q/*kϦsurkך)rnoI&#r!ݙꚮn\]/< 톝oͶA+Wz7mx?jݍ/-߻r|(E|[^{z w4,(grWw4_mdCg q5q0eHMk /dcQxMxY I~T!9łJdv*T#TLcSؗnK>nA3d@REØN24d欭B Ơ+|fLEMbN SR+[3V륅"Sh8=C\z6֝Lg)"SAvb9%z696̹ 3 :3Ԟ3vLjVjA5U>\20JeQP(7U=~Y_mMW|[5Wټz]ʬcj)]UP^Oҫ:ս_}Q[[^Wzիձ!jp5oa%Vxsu= ^{eWëXş[wI.Ç~Nj]kw/[ n\Hw~#!euȠNEBуiyZ 7,~و^I4 t"XSM3=U.ݜ@! Z,)MB*g:#8!"v>{㪕ʅ^\/X_Іܬa5MہPmWauUhȎ[|eR{M<$V`'.cprNU&EՉktǷ'D*⣯24V|XKswrDmN\'vVnV =7B'4<;gZ*֙bY؊tamU^ꇄ򚋟:{4߱MfQ::WC7[Vm-rj^{:D\&Ν})k2 U\T[ir FL rE Q8gQHCOM 6HСB%~mK-};{Fp%hPBKJv"%KZhx%AIN) h"H.lWGZP:am ,f,*ȉ9eI{[cÇȐ;ʖ貊`jW/vb-c1ɾr jHs Hs.jglN[:T0،hfbo{Z8"qM[/2Wy`7@'`-5a*DMkm: a[4iAz%Kq1B'.U B(i%&r)rd.Ay.G`sdD<)h:+JA?1R9#bPD)u ȱ&k7{49&eL3zTLnz#,DTYXay a9r 6YGN* cB΄d5sEAq!@dIzeXMnKєkΣO/@wڤ$[3a s-z7̭ o3eɫrhEFA*[w;Iruuk\<ܛCʍ V퍏cʿnFzҮ8pVXoU.} } +,MNѪtw8ʘHs%H ؙȱYلֶFeަ0VTeEbyCkBL-D8Anu}'&ӛI(}96;e{ܬH'VWT0;Uf 2 yw}Xn6tt'C==*ނW#ɠ*|ZdcC<q*ѬCd ]KF ާ@{?{*OS1PJ&/3AFcs\) bA9'RXvJB9-,|2ӫ୍!! CF`bΣ~[wvۯ+/$)n83߹ԑd|Y9ye`j'JZO{0{q R=W='#. `ځLpS6)ݎ=4Ŭ:1b{|fOk{5<.3A?+mH_6y0>/ƃxM.BcdxV߷zfxJCQȢl&%quuuub V痒ڇ(58G?YhBsX5!#V}`F|co<^m#?'vcuwC) ~,NJHjpˬ |,%AE/G$x*PބFyfCR {c K{4#5(D>$HT.Hm&fbL#"W+\}Qܬ;Ϛ%4FO_lQxp_pX 緗n(W;xlxb_Y0)zp+ِ"OH11 *JiT(r0~T<(hrQ5Ϸy!&(E{1P@YR1 `OQ1Ert!Ee?_K:F)g9xEy?"|evnГf)>x⁅$#<&@M$JL=PXD3!#C'J3U,$m@xUx4Z(@޾E>*P(~fa݆!¡C+3,GRjTJKBU*M%Dsi!Ath8A&#BEUF22&H\#`*QmJ }m-ny%DPAQG 7Bja}XJ >b9W G5ICc`9+70Z<1ƫbhKER_%/OOM.+U0;LKJ1rU+rt.[8(>E[q&$RʘE<2\svJω__TsbnEj_ ۟On/! =IF{qsOMݰnDs7>Q*>[Aw5ewip4^glq1=uLFr篃ϯIT3zBAu6qM y˺q dǿg^UޟQ^pDQI{du/C׼Y\6]6G]O66 S/-7]?quE?W':-f q"̾R_)wׯ(*UPjN COH>%%E~ʃS~t4Cʥ05VD<1"h.7UہJCH"<&v7$.(¢:@2^*tb8m׎4t?1Fr #'&!Dq{$A䛯6qZ;'uI$F<^1T;POh`F}A猫cUWfˌ+sB(A +tysG6!$HЙ6LPz:EVeb%fV1* LӢ!W_+q&ˏY;8%3є]M.?qT(q k#; $᜻X#;Ɔwl :Q*#Ϳ[:zYۢ7Zs9]i=%Z6UkXҜ;@ޱҎG ^1Ta2ce -cjL(ZϨ=I,!eT2ms} y\:=42^ E/P፧:c\k\ "8O` yj5r)!Uڱ|<5ǥ.=ޜrs 9eSź]1ɿM͜G\bՁe'>8QK۫2vPm?z2'xWy~lzq#,:h~ 9~S+ lyum0zH8}Q]. GUB]lWENK! T%-5z[}T+~QO'Ŧ:r2{ZTߙLGq%A훢8. e+._OW<}vRFLGnQm ODND ty`NEjq,{9c9o]>|q=tCons6כ+n#zd*P#B0vaUSfhLÄE~}j[0U] TPĖz(I'UѪ9qOz@]Wj?vNwIOGq?r {Q+=S; *q2Տ$4!:Y 90'UX:vgA쿎(.rnr릫涕&y>-OgMՇl} 0u; -F g.-o6O]-[[tWzoPJ14ۚGqO%sb)9~8]&d8uƓh9$7RN I]x; -z֜^{sZ qS~Eak`)Kpkо抻FSOȊV7XUQ@@ErĬZAqP!5敖]|n?<5ߦe1[aq²#r)߻8-3SR` Q HN\KHR[*=e _ycN>!j[v^>pasY[c~eF.XɒC NN!ݠ3VptZ"8Pt KWYN CA!Cs!tYqDR\]M, wfu.R!m| Jj|ąF܆J$eAI3c5F\XA-NP\wG}iGdس>h >yI*kjIQs`RUV'^o6@TyHt $ ۘhJW$n4IRK:Z=HmQF"Z#yR9XV[q4h}sJ> D`ʈ6mĢEFyt o"V PSI,\䐠GNe"zt4o@)IB9sp𣄾Ђ!\=Qb! dSiT28S8 q Ye+nwMĢ .k {\8pjĵ X 8.8!ʈznj+P|>.m$F]h CHΑ|2)G91L;ЃbQqDMyD,ѪMb-U DGdWJ5˵2MD6RcKMb۸q#kt 5oף2\P6t.8gF{Q%n irt$,$ g-BEjP Isev^k}sLCXY#UiB)ޕq#ٿRЗ6[kò1ׂ+ZUWtU[,cGS.+CRf6$/x$G_ RTQ6%_")(ЙVNecD/FQrX3Ir L VJ (Uv a*3jƀ(+`UjОUx{h)B6 a z"R++ክetiY/CCE]% FK&T p$f^,.zVk(@H*9R/DPPR )l,x@Ht /\>$Dk])4 tj5XKVnz1B/Yf9I`1T/cB;L'&Ի 'Cv*҇u\W%^E Hɭ -UzTu☌8;`TXzaBxPr Ұp1!ȄO W$2V** S=VYѰv/.Dl=m\@VǃVp e(mۄETgES^ DU;͊vj[MⲒAw24_Xim{u]w~26:}u6F87=ob8AbC8*f7K|Y]P6BQE\1m6&#(ڽEX|P/#m6ʡЎ2*vP 8@tЫP%l@}ª ڜ׃bF4F,%Bim҈y CPH]@Y]c_XxH8K.F|d@RP0;H-2bEXD 걀- c:¡gQ ‰TI~. =,f5*)8L"*FhSrV-u-=$=(,W0R^ ͨM:ɫP1sɺ6Z4f B)ۦw9552 f -tӭ2k& n[=PXnVn IԽ<m jFCo `Dy6 EhhCzEm堓^.Ro A(<" U@%FOp'Tz[Ոz5a2'˅¯8U.D UMIAs-cgîA6yQ!D uMga TrG^wPkH&=/,~ 33!YV2#С m@[Dbi;7G |SdF)l"q˚sAg0`~"`_XfP 9$abR竈`ebwJJJI t!ĒC;t\N"Kؔ)5P] S q,^aՀ]zDlgI+Bh%}lBImG ='`/a5@OިU[`a!1ձ+p]1q5?)'511B;w:J&gTcc}!&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@zL ͆c]O*hj/GO~T.&ׯ꯿@S)y̻tqw"捸'YG4NfUgX)|2.rFn~Փ]TS(8͟7%m泋T]6jgÀvy6@ju+ o T [u:t`?(da%}gyiE,7j$L\W塚 =Lʋ.&eБf9TQ6/4[ Dg% x] Wsz1m ۈ49ٴ(-Z dL0&wFk[ġ:p6*1y9mqei?˸dS6#[WoY\iƵMfR^#HF># Y%TsFE/lZ/5kNPDuNcbO51Syed0ݠA ?a<>*4&|7 cA#FeƬnu(QP>Frs8WƋ7a_Cz}#vXԢV\b;Y?7'sjKxl5ρc xMcKd[sۓk*y׫ܪ>U\DT{A1e@ƢkV&ד|)Nڜ)3( IR)fYO5֟c71+|6OȂt:u[uA;\Q2 h3sNdirOWUȏe\$,iCVԑu^fV$hea&]"=.q; .~d?~ /5zU=/oc3%{RLviF}"m~e ʯ(ܵAk97/.X]a)ᎰާÉtE~"E9?O(oc(v84l;N_wGk<^L&7oJI``gwڻ7e:Z ISHGYAO)G~96LNg僚_1Ҏ<)vۭz:Xz6_NkܭC.-9^\lrqZ~i}cXpp$`άZ}8Һ׻g|Aoҿ&˽d1u-7ǻ2ͫ)= UvGV jlFFH>ir#5냃:|Bc|?*h(3Fe%o,!ZUn5pՏX6`a3r Ksw5|<\c7὇n0lg`6 v֩z}QRm=ϦM*3="\ܷzbBFxuO7ĉa7b ¼ d- hj4\S|{KoanOlvs^%uv Ue..MVГu@km l}=Xun[IT&I253[Ei؉ԷOYR2dD|6)Y܀ߦܬԟ;\!6-QOrE#/6'@.]Yr^`IFnWk`5 ~~R˳3Mev}M<\*dj ʮWMq;y]U2ɇFL|zV|Y€FcaVQĬ+=%7@ӫn1xUբ2],/M}VjC-Fg3$^B~۪={]_=jjZd~ U˿il6M;/xR_mT9ʾ]j̈vu?-|׽bf.8.aN񴝌_Y۾Y(MXt߄s'eb&L&o鴯_fWY^`"2u6 =X{ztami[`׺nd_muBMQsdyѷ.bՎS~G5;V B tea(@_߾ׯ߽p} z`QRM[%aZCUUKZUS=0}臋_O4gg5N[+u;(L(bc.(TqFt;!օtlÛѥJAYYW1c-_F|a]I֩zO,wQ" 2Q|pYޱd2V WB` cq }nǸFR:k.qgEnF\o%JòÅ|7 ,¶hVS҅ن=շs42N\Za-Yk佱Ķ;1{VW]n@q Wݻ*>}L}Dc2&z<O"AnrdLWJ  eCӐ`4m6K'O<_d߿;*pm]QX(& kD6")br1gY%Xwmm:VOx/U[<I'앜x2~[$YiYv"\0[̬ _&אx-9gcn9lɋшLW1j)6MhYGRRU-m9;9r~wp&ERdp9?\=^ vʬC\hH_No,C4N?ˆO4)a(*XvP(pȑuHLXs:cyԮ K#tb )2b*%XLtƸmEm9OkZx~qqU7Uo?{5)FWL%zfKg3]NLy/QU*{rV\2DQ~Iށ1Xc1*}ѻAUUϒcGEYGd6hd覺*Jh;YgW<>`Aܨ TX\\XL:[AV/bJţ +j.F=U{N\X55xo1zE%UZyj"E=&ē&zFs&R2VFf'+D֋5Q5%LH*Pm8{ GqTn2Njo.t,JY$԰hAP`NJG#2I)K^uX'mj<:ȃ^u=~f\+޿z0iX*Wk5!,:WJd''N&9N aRq Ɔrqhv e6֒?d0CC:$Z*!ayh|2B7 nr=DlsjG& Yװ8&3D4tQ)_?\Xs,KkcBI2 T($Tȴ eQ5HkL+ ̉ٻj9͔AL%CնN5' wFݖI&.|}YQAJ̵5$3 S *) Ӡ~ס 9VGʠօ[{р]b"q2Eր]HN hT,rvMKyhA;*rMN!w]ͿU6D1f]djLDHO>Schlo^ZOÎl>Ǡ 5Y25lVYz91BZt>JTsV%"bz[fC) )0ymAbz[ny2*la7ζMEmڵ]}ixC|LӮŷEY6_Xo lq-6bjII #jt+WR9_ltBFmp ?n KV^"!9kٴ>p;@pU%޵z-i/ftbM^ұ`Tkrx Y-/IXQ~;Jb TI)R%5=۫-L] 2 ,XbkUM sf#6JH#%j9!ny+g0 %Z""NqkRMHQblq$@AؙAhvEd'Q.k}OaJ5=;RSɔKuK&m dYFLgm9O[SG..rU"N/).bg'8ŭG(Z mO\DP.V֢$6. F(Y@CM1Zd_.=&b̙v[h٣o+<;90[xzd$)ۿp/aX;*KsyiC=@r!c*6CŬ(C VS"ԟdOrIAĮzFCZ)#iR1 H怭 *_l%߷xTrzscķ؝ٿ? pZ7o^}:=Nlkmj֨tQZTɰ5S8oaZ#s߲ZOA[,Q[n `EƻdDGY;`*NwnyZoQEK\wM;az)?';n(B?:Q7S-!gne⮯]ADk'̴hӲbh@Kna\㠙bԬ35Zs,tϵg'!s+93J0k) n艄9 *XV-')!l.MG~^їr)3Ŝr(! >'oxҬ VxPIHݐmw;Q-)_Un -.|_mL6֓a16ehڠZHd~B,]×ˏDPAU:p<(3g#_V:6O~0/ܴ7?c.Oa3"S9jG Cvi >m| Pʭ%NM 1 Iu+CH.;k.t |*Ev4vD(w8z^iA6:?f:kg4r_|)n!MٱMn*| ?q F\?q~En]BTw,[3\FkOfEfgS/:}xуnFo^nlĤc.Vng-uif̾~fk6XPQvƈCIdkB) X7"]"b-:F7Wa[| |hܶ0rK(@*% `[WFS6t"%iR.ʻDo$wiy5#ڐpR~jcFŎf MO~;EiѼYjpvyդg1s/<|N{.AA?I!lUwE-Of_Vj¡YM8I1* ! ްh*²C@ -DticKbbRչj@}k' dEg1RMA~LAu[:&g‡Hт->7>/|8G~6/}F`qq+X_{kys.T pHhUVhs` up.Pk'M'449aPq-X\]|R~=m` j}AݜvK4Ifab]Qy5یĢӵjRT;h^Ô?%>~yAus՗s|ֽHgcU(eO 1rr&Oqvӣy654mHjk_Q@û2aC{Me=M>XܫC~k>|ן7DŽ l>S[6VoD \mZ;{%PJ#epUlHEL1 c'؏fxSu g PZlթfIhںUf1S59}ό). 2ETJ']mF^H^EA^JeM1k9O;sCSɒj1y`jP`#ZTXV??W~=9NБ_5>DžVdsQpPr~hX?l;W=puTt^U_"p!k1[\)9A~bG&T+89DQa)!q ײ׸)6޳J?,jv_ݏ7]29DXٻ8#WZb!N8; TwWSRczfǐPZkLoOMwU]Uaɵ 8de\(n%OҪh=0A>x!Te=$~aZCQ#ꓔ-qE͉H_ċwnvA7z_93 ́6HkQ!f%dJY#AZy. Wmۤ lMl=UxMN 2lz:jz^HP}D-^L}OEtOx űzu3sY~qɰ_!md A/0E ϧ݇f}\ԬU%+n4\5+qB0!zjirX}Y]e_"7Y1MZSD<6uM'MZCo }(k׻wox[ OZ病 ۂ|юWGil9Ο'És ?">&Lo뛑 "vAٵ9ZgRNH*?OߟA_'dV= v_jYn<)yQoaZMRIdU1NLӼ@S92l Nфi˴K1 * |X*^sՠb$OLIIۢTzGigJ/UѻU{vۻۛtwO.]oC"`H YYT&GiWx)9$j̄(| ʄ`I*Kdp>\RIHo9zgI eFΆ 4nή8ƊI㒒ͼFDDo88gE`Z jӼ""3wЊ-z҅B1 uldd sA'leLXĜp({ NB^-*1F%)$cWߠI3>ANqI@Er #0/r@Kbu2|/c;$cs2Is_=~® W׽n>mp#_ƑY8ĺXn @x'Fu@V9Gc )w F:9ށS6&Ʉ2 !h!Zf20^ ddrF)rF%7b1MX[8^ BJ&V3LWpǟbx1VۧDћLmMOU̎"$yUf\ 5 he{k:nj層yE ]ܭ!)aIJ=U+z1oڠEj{]Y&}8<~ $=il>}?NZLP&,?ˋ`aG%r*Cfk-3;n!YGU(GYh{0q#))K^钌*SfϧM&hɐ&iG:%9R^(xaGB'<|e:c0diw vwMHfu$v líO?il11uB/2i@cgzv<:/ߍwȆ/0r۸4^|0j{\jr0*ݚ;1L6@I,ZcWY`nVOZޓ睇.40B1@ИBtb^[Wm Yt /JͧS)Jvz5K mՇ1~MʍPNozrЉϒ/`m?evUz@>׷̬ߵszՖMPLi8^@DA$sb';)tε`g`+"of~6m*xO6:eUݢo/[% w{adUcs;|:W{&sWulҵ߷M .B8@$7+#Y@ƽ8&r*E=ˠy$螒k.H1{~,CGI rG<&y.Kt+J:iT0As`Z*`rBĈ.ECɒrR2HֶrJBI$E)lY2}rzo^Rsv` ]RL8e""qp=bY!E ϕB.5R9V$sx RKCȞS9NJ>^Y똂f8gV%5\@FFޓLkIbwR~[$4YÉI:Oc#C9@*36YS={8P^O֗y[zwCs@3$n ;W_R%i[k/]GZ'O7^x-Ð "`Zs9Dfeeƣ$7$Zp4 zΉwJduZ-@c-IQɴz=$CDHJ,7\.Ft% GC%XfYr Y#g$3` q3ze :,h :2v@N&x iuIgZ=! *.XZ/%Ɩ^2^RIF Ӝ&;'AZG8nAx`ak:9t)ng$kN 噴#MT 4iyīڄd]$ yK4ؚmNOrΐngѝr,F%GhJyQ Uu}YA܋Um^!m-Ia.ymI^0C:kj:o1k<ڭF0:OiJi״ursMLKijtuWs-uӭ2h&x ԝQ7)xv`)jYxZ\)8)  j31`˳Pl Vio: <%2`:39^PnxВ}8(DKh wzB # R*|Dz;)=ԃcFշݬWa]IQ!|_=oa+@1\ƒ urprӬ:l3@Q ;b}CQwwmBaT(:#> ݟAUݦ\1#dԆE=+NF"5R&XQ7VVX/6'ELA I *tTSEcP?!\,Y#9v\[ἨSMg TvG^}F7PX33¡FXV2С m@XD}/ĺ0""U2VkO^tQA,2 8ÓJ IK DE|V;V1rI"VVuӥW1ebBv"J*I ID ?tĒN 踜F LBF .QBqUҽƂ;]UBpюkqht .>6ȣ@aBƜZ*Q5 P=MSp6(So}[x#UZ1y@Jjd*v(JjVMJR3GJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%U^R)R7y`@֚WJoH @2"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R=[%L'lK T͵P@VWU+#%sT -$%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R*8$% 6p@0׹Q~y@RYR=G%hR@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)H DJ R@"%)hyi*}Ku0|{QVw hw*%'itsqp*.\U$ qqCbb8"L|֭E}:v|N.4>}ޗnRitܮ9Ts}uy> hs  zWn*phaE|78O-)l+ՓlmIeC)ƦE4JxUɾd&x +q:*&0$0'Ұ~'AV6hKR0yL0Yg݆ aJܩF A" ?1/#)X;w>zS{ַ 4ڦ.[(քRQED݆g-nۢ결p6*1ߠ`8DkefngP8X]O$gkWvճa|5 ׌oHƜڿBL(I SF^6R2ȝVѬT&{9QիW_};G WwOU7P~&,TiHnfY:joF-܇R5I0kCay9o;C͚yt30 g5kkT"gkvG6jb3(]3tR~3$w"|([32a3D Y7/l w;h|ܾfV-c>8%+uWKQZ[T)m1LşG}۠М+(~.mTp0~Cc۾mFm;Pj :X/Gf?^~lǝdlͱh8T %AƖrɶ/%H~寜Cox8받wU{xw}-WSY=H:{MhBl)7I'8{1Wrߋ_spϸ{+ٖxaC?13xW$tI0M߀S/Ӌg4o..p5pɶooΏ$8{PN 0y?YOxDL>Mu%ztO4V!$DĨN. λp5끕drd1fZsF1GK:(w=wo 9%ok}Ϙ :S6H#njjAwۓKqH{gQaӳQFW]`eEgښը|uQ_jߨmW!@1mvj֓ꂋG:{w1|]MZ iQxiѤ^ !lx_Խ¤1|cb0g~ɯ&i̎lMݹ8l+n&ƈo}9vvY q |%9PRsi#՛/35J { ~2n(`̈́8qW1o֧?NG!I鉩JH-ޱ.KPkfVHoY7x1<]vPCZH~a8*p`Qzbu+³<l~\L 9WO}|>}j?}ޓ`[ޗf|=I9`UTyưɥgJ\>V vJx„xtV+ujB{s{WHp!miޓɻiC1bWlYGp_sFkӤ:uxA )-;eq|ͽ9{J~iG+T%rut89m<]yrˊecv~ ^?Zvkl>ʅqF%dGrIss0&kQ՝>|99}u|НNǵY9魟h$qiړ}덴 RQ]6dq0\vu| yvzxѝ'䍎MQ}D"?aoBqg%+U3˛:&GMIޠ 7ex4mL-Z+Ճ* e*`4µ"=l%=y*0? |riHvGu,'ɧn5&snԓv9ڗo^Ýa ݒW%$jxzfW*t޿]:H?Fg7e)?N>emQ3"2q6 ypۋޔՅsޣpQ"r 7Hb8ktN'/9r2z\|q`\'iq<>zypƅV=$Z8tQqҵuyNf9yt:31`&ߜw?d<-~>I=z4:uN/r4}̃4KHO> 3k|FgˮrUkv/E%= ToSra'[w;)/^ʺ}‚g^)o/T"~VO w9'Ons!лQmՍ"w+o_XRk2<ϔO=_lDޝ9A%}֘lMciC2[ܚzh3:Fڛ\ҿ|Q^#o~)\˾5Zpط{Me:\\icx!DmkS9h wILCAluJ(er9uf+Zc5u@R& g2·h :h{D+ HYӵv;#:{W9UWܼp!n\U[#kHMW;5uAذ/, >;)m\Tc%D' y*2AY.kIƌ6D!N2QVbot|@Br_&`4uշyR|R>L:>;T ,ebX6Gr#;)!sսIl0ێl3dj2p 19l}܈VwdPٻe{JLHbUXroA/fN[؛!S\ucYX&oe# e_viDN"Z/m*9~~M1YX #Q ,Ud$DV7 .M𒲦vK*;FIi?ޓIA.9`,wJXL(Q\% 3Nq$ɹTJhYθXJk["# 0 /b2q9?5[94ϣ>'ќ_'0IJDND\Q,n>)0dSLt˜ E\2J`5Ͽ6&=&GF!2闛:>+5=V>/}%Eelfyॹ ˻xޏJsˍR?U T_2qrP*6 ceiwclRVv[1%hV-.k'L}쮿&*SjԪ1>{Upƾ=*c(jykm,(Vu:f~4€&u=xm5 16{mDYۡFAP{(.Klq__rwCWZD_.rL zy42]P)3݌7A:PymXK_Jl尖 /6x\Y0/k6E'8`Ze"d̤ r,ύyBnccꘀ#Hd e.eKsY_OmV&tVz(eϐES C0M'e+]2qaQaU/CIQ./ S9ïDzda:K34~[1=(!戸@2K4K S3q};PY 5Gۈ#SkX;CQ)3,,1^eL<)!ZX<VnEl#T+s@ܘaGZp %$%FBwcmA.'[o5W Ӫۊo'LQLm'j%^j,gr/A(D^A>,Qa? G-r A{6`62g(&CJu#**w9 m93|}Y-gUxد6e&-g02ftϲŹBFA}54!'k,Ёe{X+SUZjٮL^{z>ڬDքVCe,<%hqGKTvai`1,{[+$Ln;/TnW6M͇}4kP`2,苿,;YyqOfQ_E&>:,̲דU]k۬0қ I#:],;r+dT2ތrK٬Mj/9eYR3ZU%5,Txac/-m{U?/VbNXTb2B8eqq,JBDx`P4r*Dc ]U ;] ;C6#iU!tp׈ ѹn;ԍ)Pҡ!ePƁiF-|%Yٍ΄ϑys`hT4%~.oSM>clnj#L& \Q(Ml:esZhf4À޴6Bڊ",11aDbbEC)P^+{WN û7㜓dnZI*T $v)[${2QV%&B0Mf0ʪƛmTbTM_o]pU,è4+-)1!zOMhDn!ȥ.z.Vt6V88֛qoCI45̲)@DW)u: <q 養y\tB N)e^ BZ[4D+w Lͥ'Is\xW72aq!QB |H4B|OYc\:f$EK)eMC|X+$ 9is^fP%no8" s_Q ЫhQ%h M[S*ƥ vGs@/ 9> <<ơ4JxN2ƀ|X^땭i#jt[iObǾ@K+T] %ٶN^kQs&#֯l2Mߪ Q*/YT l1oUni%?ܔapͧ1};dn~c=g*JHt@'E4T1%X+`R_nJ6(vNcC7z$'i =(Jl4ybEcBwyḛ4Ky4V ELxFGޱG0F5my5یhL\H{2Zi#Bo\{aS؆h˂sp+0qvܳ{ ڲL_ k1Z<-͇P4MǙg"xI' l?J(| UIRc.G1&J }q&9>rTqh\4De͹j 䏾ޮ_,CₕXRRRUs@ڬ{MN[ Aڻ FRsŀ\>;R]۽$[6@_V틓れЊUZ qQS`c~\Zj/oj09`X; mkA,h F ^bp1ޒ6=X0~h*zA;&8~Pu(Ҽh!l,t^kykK=

Rgp&x_' hG[hcV$.^ДٳnU+ r sĕ2Z+Dɽ\;4`?vИK٩~XF0Ԯ54X @ sxeFH xS vBHvPV8*>C9b邰pΞY.rK.K *|1:MRG4pA@A09Vjz9Bi{ _^u6ڻJ ^IŸGsOC N+nHJ~8l(^a[.e--gy:fk5]Lʏh}/}4Zy xdb`Em\-|o\S 1$*FӇVNX WՏ1{ryDNӁ+hKc$2սcQ"'"s(p:s ڵNa B婳d92[_ Le1Ci|8s7T\r ),j) և <)U[ 2hokǔ:!Cj]<3[z!U~ou[g:& co]pqŪRgag#[Օi0ov* -Hm-e>*mO}.00ԁS6f,mdzʟݐT-+t.oS"He}YłˊG|@ $;Lp[RIA,y`]9`h޳y&j1J<,d۷~S8Eل Pn)BN֑$$k`]lx(>2XR>9O:c*oR`[#[3d VDxWFY.GDh8tQfvXIiJ6 &ŧA_!KDK{̊GҖaky}h>'o٪giba\>|w 1Bx+ec.XqÕh.I!SYx\ՋޏlZ`ФFۿg?柖OFf@ Ohʭl缪h4 .\x ~fr+%՘vLf/IPWbkE.YKV\V#T:ZJ n u`B.j~lw]؅Y \+%"WɆEc.22rނVJ[AJ2NʊcBZtIΉ zv}zWZb|Y~캬"BIMVwҮ]pwlD/BoJ^j >]YĉPߞy-gOe[Ï<=p-/ض 8+ 㷪9jgvKJElz~$Ѭސ8ڰGcQKz02|FlO_ =8 DvK+ص[&B]qYkGSZmHH' B\4EbUK2B!{-j̞M Cڬ20w 須={f-1wG>G8H.q!L@].9-zbq bHZרgj ~蹴"C\C3W XE'ߟ f!ڜJ}ah\:. N8X]o#9rW ]I>Cw[zXy%-kVzX54B|-8Dwe9v֒k؀a"5Z7_WͿPfSn#1HԫGF:wCš3X m*P;pC0DMw2:РUtK><>RZO()[}J|ϑЕgkFuc%Bof0@3H")&i^mr}1SoB$C3@!uq 2]f;R/bJʦ6}_ !x.1b|s\htSj*m$(nrD[ء8fXqBUw\Q ;"oєaL s(TfKaR񨢼jw~tTJ 6 F.j_YJ n "{ >Krb'k6F`F (D=CC8PpYICgtI,\r֯![lao7@gyXzYrƏ#BO5ڻp>YO[kdl\wk]%T1T>mK+p$g]igW'm/uR|&^NC?éG r~?觙l{z_ϛWa#M3/:t^,Pf%PJ`ty9c pW()欯ܔS-sS^JcQ{[zws?SbljNs]qRD°q7*gzRW VD;`L6*랜[g+۱uv) /߁|QOP "Fo)" u^S( V;YV hAVŎ~p H#-?⭾)Kuf>{*8ʕ"hvqz]Tekv M rS5Uw'8>ձu Ug\&L֘~^~|gyL3Ogtq`OHMC.T2ĴsSW 4W-(:7;P1ue&mWH~Eߖ|ُ}^:%+Hf3<28XxꕳWՅhB^0Yhm?YZ׌q:?S-L}Լ)D?N?83"{x.d>B~xp+M/_oCa(N8¹Nȡ-;COƆ/q.0E Aߔ׳z8T\tHC m/2%eb0SCwSVp;SwamÏnF>fp~Bh"lhAЦ-oj^X_ǓwAMy홐{O}]6BN*޷$ 4zȃ9PvD^7LKo3K |=\Py@K.ISإ44v*`1TF͛Xz>wiR'+.^dm"gRRA0\cb^@!=Wpggh`h,8 }(UyIp #\¼^r bÎ{MGil!4Y]YD+EA˵\s׵.8;J߲} KjV6G4:=.BQs77Ac7 2Zd];"񙛗xPPPp Gr=NZf#X--՞%ynnEATFLrpfs\ oSA5~y=uf`5E \C.tG ςk87UaA~ĂROy1H%z}W\[hCHSߞ&V8T(/1yݨfSX'afkea6Yb&CphK8`)iCUTC_W: *MQ}h:+q8mzrhDdԇFB'qhf0z|<#y^:`v`dMwk])zs!iLzc(T08P-5@֗/r2[P )VF1zl`&O^FM IpߒM%{(D]VVݿԫ,|@E/%MbIOLsdUE0)Omr>o+' >@wz =iPTV7uW `>j LGxs"l$1/{z`@RULl'!ϲOV%v2[5Y]?Lm.ԓvc{収{|\)g[o!x'v8z87/ Uww#9|T>B䐶XN@C~p[R ٲoGt2ֈuNu֎p=Ze2ULsN`rLx/OFkRhQgU< 2>@Kp||ESFcحȻ=2+ LU!BS&mRjԱK Q"Oh(B;nQ5&✅F^Y幦)UeS=>Q)Q.N9U)`c=-yD%e.n9g> r;^' R -toT@ì&0.ϡ~ztMpe1e ;M2,bܝ6ɱ+ MڵZ(}NV|iN)rPLh ?0c ^p\IǜSþ8w e gNt?[7]GzN-[n@Wײ͓Wcz7Mm_?ͧr2_NTeVM,V_F/&uK/Fhfj?s_IY~M֖)}nbJCIK=N8Yn4QEN3['C`Ѓ2q \q)Hs&Sw J pByOo$B~'0 ϑ8!CLBq.]Ob"?Q?! QȊP. "wºj|[HiJ_@S2 Ofڪ5SQ6Kv4[3x"{׃M[zv-u6R؁=L7^ձȊA ÈoͳfE.sj):D⊶,e^MuV/;\PnZ,4z"a0Kw[f.NiW*Z޻պ֨XGVT[-2Rko L*kfCۀ /Z67j>+Gm,?6>=ACZ:W9]4ߑ,$m)f똸ii+)0P9)Q!J%ƘTa@rMZjvȡtԱDBJ^ΌQcYQTBt& +eP/0zV3WKÝvY Rݫp`m<=ȺY0c-+ySk_PٞSEc^xF C Y/ߢc+/-?l xY#Y@4! B+B)u5vCU6] υ1v_Yo\}ݯ;yS 'ͱ&ʬH[{Z] j4%,α~{^]EVnP|8bOI6T ǜz]zrOz<oXٻ޶,W bf,0bg0<,zEJV,˶$[PeJ"(R2FY_];tM/ᄵ92i|ȝ6p6d-ɊzKV`qa2:# 6U%6.N.%d,Ij^ΎفӢx[1mӂ8圁 -@!IIƒ* m:|_Ⱥ4dt'D'y0%,.x5~ոmN%7^q1h2Fn97CtKysӇcRɏ9gA@=^`Ɋ -b%22#GCBec5E{`НrIbG8i(L!SW=eF>RˆǗo9ByWgZ4gyF㟃}C=0ۤq v>;^ L݋Ræ?pJt~n7ғ>;Ílyŏo#4(jQ""}O 3ZM]_O w)iJ{se˭ M3vm#\NrDTh`N|!\b"0,*7paݴ { nK,l;#Ztö W$ 3(|N`Ң"򚉚xȌổw+p9ʡ3lOGfN^F H,%2.uM {wm| ¯f4Nu'6iGY!XB,1BAr//SPIOD;Z=υp$h%z~Y0/ԕ/IB"CAl d^ }GQ.,+.=kkIb^nc^G|MSmeF>G}8>*^GS5Vv)ܬIF {΍@ܧ[ :PD),MGb7_ ~C4osֵy?_|?֞q˽^Ȟi~{ n-FS;In" D|ˏw:->#FpJW>=k{BQ̚3CG43vBbM)Ǟ1T-bS+{3H9١eSoz+::q1ώ'if;Ll '7Q?-|-";B!D+;Rj/Cp-m%#CyNe =pׁv1O:PfTy2@$1(\p*ր= -Gc7i:cK8 c:[\#Ԃ,/)bo|)z?q]9h:>`VzSB*ʰAgGwNRrN-^g/Ns"-$n໸@COa)b 7KJ j ˃jEUS1xu"uqkJ!*i[Br̀!cJ0".@y69Q1*rurz*vj K#aB8imޯ[5#*tRKd-htDk:ДDt/lT^1WwSG%6Se 7xČI4=}2rkQ2UǢE7 " #{KR=3On$ *c:v$ڂ݋]Fe`43,8I)C57r蕀0x ;LLL=Tv:[.Q7l eRR"\V<(QKS9ȧs/k|e2`Bڬ4zHjocv5ۇ^J/̷pW9U:s眊Ɇ*cПFZFSi܇qÉHmKTЎ{)ZJAVHbI=\J/QLGqy G"!Iz !myzjҒ%viPI@QcpT\Pzuns'ⱿgjO3W $7Y#1fӺWaGl^%TKe !7 Kg}C!u" #3õ]*wɯM@.Qk3 oݺ&{Ckĝg9蝘b%&LIoQ$g^,9CrBdփQV,-qU*mWiuitTƂٗMxKM%]Z:bJ$K8h풠K3Mp,|sl-8$M'-[UzxVBªFBuJS$AKӜV|1bE,MjXp Cl,RkEHSKʖtԌT&mN17Z$*E>ݬJILp{Q81ƶzulIaX?")Ϳ_aեW).=*A`|7UJ߄xKdt(o+BhTS4K&.*[N{H; D`t8xl6>X1ʉЅH$\ ~* ['R eMBI=HEN9CEixfqȬ%7[LiQ56^ mR BaO宧4M:% 䐰V1ftxxIU)NX <? AaU6V-&i} @ARdcN pP bݤ˕cWmeݦir`.rn M-1`Ϊ7t5-n8ܳ9Y)ӥFLJ#R& -U"3 -]s?*ab0\a 2E>U͢K 0t8e)J(=f_[=W.2u.C_V`B,2N (LKGnwM_DF|6ffa5Rj}"ue=\ \وSE%45B9[ZgYQHPQFۇ"`+{]N鵠i|9u6QjY0~D u0|3F9 .A\QC$A.K n,~ڥUB\KJmƖ]f%f1 &:cڪU(ԍD啘4V{N:XiS|?_X|s@KӫТfę=۹*tg4A*X1a & jaR2p1!˵MYWP 9ag8;՝<\/{aJG:(=An ,6oZP}]}Ӻ\ T0BIbsWԣ*euJ1&-,-v!# 1=$R cs+&-J$!apPE69xSY6JĂ2nv,PV34ԴDx CV|9iJ7;^Ͻzf9 yH XCm2CA)̖Vլ*SVUh\$kE3ge.'Xg a%2 }!ceH4MR󒧚gh?U.s@% KI2OK9ri֝Ʊ:9t.?"FK?F{2GtwIDS ޘdƔ0`_>Ew?Dȥk&?OuD?F_Fu_?LX lqu#U.8v | =20Bz-oBVFҨ2:I10J#o ]z}u 0a!Srގ% }kRaEA͍(T3 }N"_qWN0%9ֹ !6axjn)E8l !+crB1.KF#+-l@Uh73Q>rQ=6㙉SظLw1"M Nws~ $|{t9~ݧ_@a]~-Cֈ Opa_m'_/(EDvZQDa=1?Es#a `CJ L%V-)!F~J>L)h8( t7SRʖȻ$m:p%YyicjrC=!Dn 훗t̩CQ& k10z&M;npܥώJs@۾0 Oq2o@s_w?||PW#̞zȳٍcūs! T~h 5%1~?rrS zc'*8D$&QA1nb>jPށj\)>bE (,A&?C|lj5z8̮*ɤQ`l|lqEoxz8xg'~x3,ߖ`Q,C"?o@zd1PMğ.6rG YL#_7~Ǜ]@պ~\ #\Ac=4L fȈTdbdcl0֣2R. { c"xm[~|Lτvαɏ>FSK1W$LǒXx ,"ḥpzy"sʣĊY)gh)F)tNi뽗ӈ kEu{<$Ψƥ64"+ vI9Ϝ诃M͏K7fYyhs6"IrSoALp 2튲w] 5uK^.)`E.^!(ZT{ľp:UMYoїה.v N0ɬֿZ|l^÷ŷrG9J<KK,eDLMh k V+5^(dr&JJ'H(ݓpC n \G4iN [ `S\8iAcNسC[0j3Ȯf\Nz7zo3Yv,<}>i\)22j4<>PНoܸe ;dOVIud1A&nU\yz4g`B ;0W ~Lx=XRwXDJ}F Ԥ+phKz`fO!."*!Ah持z6\\)V2jиMFEXV:c0iKHbxC簄f#AjϙV4kr 5z2sQu&>W;L\ "rl" f>o;!D P@:$e*Bn׼vbs/%<1V Hlˣqve9?֭[dM =+aAje]6QN#nRH!^Xֈ{N\ENO*qˡ͘0j(].Ae !O5mBeБ' &Ar$Ml*W4ɚU P #5 u(YB |)q>j* k1^_ rh%NW8d߅v߅A[oִ>W>p:H|iH٢l9\Zd#p4!:gV 4 @"$qr S@>մ _[PV&fu+0kB6H#zp6/=HpZcBɊڱuWiߦ+,ibL:eˋ@Q ?oy>3/`]݌NlNHT-R,{܍XQ $h`m_A~w1r1j cěOPk0flS]ݞ-Ĕj`Fu@OWmpڕ˙Yd[z ^J*Y\J$0G@DV;<ّ+NyE*' Gq=W@L(>T'(p0;\JE׊NV/C fn;<;R[Ƽʀ`T's5*Xoz;I k~xofݙú`R*l@ěc3R"$ 0Br:r乍FCdbmѦ+흓$§zd#=YЈݒuK}7zj]\jiH?W,q P@yy&Lix  Jjl8gfN=cGăT`[=I6w fOļ-jEKSV{Y:gkq`?[vW0OE^+v5.n+Z[x2QxUBiaJ)Yzo#Ah V)ŘPXppz/f/YMV'یܱa+Ejkh4iW, vܲzW\qQL>ueq,d*y~>I'.g!OWtxQ+ 2 7wXb#ZǿGW+܈|^扤mߞ0 KEl[tm(ku*{]oВXo-6V3i/GپeoޢN΢DgᅠT`η h㖛F+mۖts۷og1T_Wm=@2cU!K}&S&*fo>GMmNkM(V^V,i-Fõm_}iPxҲ7q֓bm瞶# 7`0a?-QNm}'&P 6ܯ2v`3Ĩt@R0 vڠPJDV!1\2_Jh4ƵD:iq>jg1 \Ap/3&vPMo`5Ϋ+[ ]k3 k'L.}jbQ]~Z ޝM{wV|ŗr!nwр9$)DT=ZCF3Jͫѯk;Jq܃|]7 ƧwHn,,o0QtoT UB!k/P61XP䥷>Y7FAiLz~^{+MR*1hm"~+ya34I0ăq1EA LD2KYKj օ厥,e^r =={KGMi>Cq8Sa~Zp{|:?կj<6{f1~ڛWUu@pM*h`oK/# #-(!D߫ZfIv5Zp-?>~GW|3ǵ@QIe9eāmzVg\'<7&@YVF^GPł1)tN-͇G.R  *=:>L,c^"8ho M79z1u V?=9+Qă7cs!7Q'иT56ԍHtn&J=:@,ife_8}Y`ÃC%XQ ⨱SVo.{775uR;$еSz)/-˾l,gevpV=I !ˑgF5%v϶˛7QUΝ}σ*,F8 {|p(T"P}W(q\ FU߅^O͈ Quľp:X60Us'F^|Mcɻ=cjrChL1 L Hቘ1[A.h,Džo )LF.踄Jsm|.;HBDXW`]3 9ܱD/|dx&42d !+PZ .4CuNu&Ԛ0Qy>|j,8pjۃQ=fqZd,<=+6ٻ޸W~*zdؗB]%E,#NjjIQ_:}tE:T&)_L62r"#-2r"#H7TF/ 93A-JyK'LΤ ;T7РPP?F!Sx*lP4/LT2 [v׭V ebr>\\od|qt܍N]|Hl[  JzE/tA ! iLdX"FZbP놁FhMMN.g;oht6P"Wq>CTw֐xmb@j1z9 5֝QK+`w-*1IjX+Ƅ*ek>{ĺx =n<vlɥB878Ax<;jdY5UFWx-%UQFH l0`zz7v٭Z> 2Iښl )dK:]]|/^!x#6 i6PZt;JcqXU- Z2] 8Bj*mWZag.+qakhg,Hı/_YjJՅ*,YcnIRA-0=Gɹ8J_OΆ[ ֑"ة?X/8v9-}SlyQXRKpWDxY^[HgKڌ sG5,E͆$B0Tm+ZZCu pNA1 ȟ/.|Ʒ,?@d I[gfwq9b(s/aJ1Y_-!IHL!-C+}K>qt|u|܃ Q,n{zhi>G614P<UKR E"v _ʃHFㅁk=7Ch s5cP \Rzy t"M/(^hICn`e{ t5`> ԝ!d3`16'̏Xc⼒8B]l9[H`/"b$tTe-EezC[mf~Db쬛Ag+{kIP k;}7HКJ,dk.҉|H\uC;)vhlsqz<tн=ю= n)2RkddɠPlSvB7"'0@+Ϧ_w}Hۄ;NbR+\du?QsLV񊉌vh\<+0Bjx1\n8mI]0bew7^J&nZpV#y9²bXݣD=Dokh\4X9\ЌsR6%sv;i%U/̏It,* #@fyCA/8:9/xb,e+CA((%Qyocޫ/J:ZMxkKYc ȐuNieMe׿n6^?&|7ٟ(]ji ~_-G L[*R}2 _n#}?'}OT>R)* <֔J^b ,٫VY)9y \=hˏo}7 XȣivHtyf+Dh-#mbFz_ؚU m^qI+lr4jVvXP&2Ga$!8w$midoLHk@69B}f%EJ-+)Y^w`oύfQ`,`rYBL'}LǯQq'^rъ:/k8tM~;kykz_otͶ[Vhٛ?&dztoOɷZ[V}`dM> gofwٷZxCYcc`;wU3mPFi`f!GOsOz;Pd7{26:KCF f7gxeSr U&rE9xL}svCZ*yL6Rvy^n)\nht])1D&vvu IhT:5S4:7РsC?ڷm]yl)c͌7)6B%hc}bNgY Jcfqm!lЖh!ht}DUzcam$ 63s"mq[E[Iz1ʆZ?Ɵ-XPԨKj"X[y*[3)Eu,%``!l/ ie_ڄ-kK2ʇ-&\Y:dLTzu\#€'b0T͊Kq Y0˦"[7L0:by5ϊbu4ćQ .b QBfrgFucɗ)eb&P2Jh6s9cBvG9[7<`סIl =Lg-0K `&d䣋2#!|1wtvTUu8쑗*0OުӌMb+ ۱˞dSf=kPF$/S6)@eO9'@cni00 )U5ԓ9L簕2cC,!" LbRL CӀΙ*q w[~l^gzȕ_ * ӽ `pgDqۉoQmYZ8qK̢cd ][`}_-?g [Rchp_#6bd,clQGkRV@J }SQ>Q\,ݠwCIG٘t59e ɍ`|61M8\'i:cu ;fv}uv 4`ѷH=gw2oM ^1WMdh\}7Oɕ{w#zw**l}P.b!Y84ѵ}c%yxh7F-ź\î. !61F`w:zuTp`|8Ib_M *Q?tM3}c SL.YC ҟv }=L/yR=X}eiT"0R|Dχ|A Nl:miLz mM7ׁ,u:'yXHNbyqLO8QuQ G<p(Do݅B9z 'hD%h߼ ӎYCL9Q"? d^ylh\ȇ Riډt]#,7u ,.Le "YOt}Ky3}k }*9hF,^9"_1-{{MkkM)H5^>3Ȃ)SGۼ_AȦoFyv |AFv(^q=vNA)h÷\t\9Ҳ?G|}/?8e="'rR9#-3g]&q3B:)[\ ЩۀaC,s:N'taygtS٧oN l.O?ZԱ8O._W<腢Is<\*hzJ٪=K/U_6u|a^<m2/k |g߯[t\U[Wbӟ^z}Kn9;Xl&rը ?ߙZ !D\{+} fB0W_V BL ߽zC<ʿܬ1}{Wl2J\) ũN thz7ctFL/<1hU ^&VS .~~l]n7J~pY\0{ufx1bxO&@ԇMF1kh#-N4-Lg6DՊTvP=Al0LkՊ0k@hPyjU,갛bT8)/\|P4]82\vrxE}cή}Qzn6c~gw  {ߺݭyZ@a˔(ti,YC,é;vۂ SQF1cܴv- :`1(Se lo-0 7~_*zϞPćŶP̢ܮ)@*肺iĖl=~ʲ~GyȺCት-^.[Jm{=bWeɌT71}cC$:v܂򉿻PԆa~̓6D`IDR&uK:ul€YF'&yt9ke\r11ܚ&o}MTHlu೫%~j:?ʼR^z KI6Lѿ)B\N6%MO ޝ/'ݗhr3C"{!l_lx<["`W{T$S@`_>: "ڔ@S6ɛB7x@8܏9X|:^n@hI1 {NV9b ) 3ǗFќF=¿ƬFq^Ԕ ~dvIuWNUl9BC^떘)8 2MlY4*A"%T Mu|KƐ sR'oGO(A+2XlWF05Z#(^$PI&AsaC{mM7Ͷ"#_(|tLJj/ߕuqCՋ^Pj-8FLοlzabrƙUz^iIqaZN|houL oUŽn^[Y:|:櫋h?ŵ_S{uv],]VA@ZT((igQA^DV`v!uU1I*q]]+ϕ'*$2%unW-$@;RPU]P"j\mm5A>ArV:Nޢ((&eđ`B!a9+y9]T֑1I%A}*CA̬. 0UZ2(j\ ;D&'-%Y.fiؙ^0w+qW8Z4V$H.kA\O70^ YE"7br+9Gb,S! l)fTbZ3̗1Տb9 ZgNFFxE>^!,%ТrRhg[qz)!;!y!o9 'ĐM* ([KĢ[֔$ѹg [r* bҶzeDB:/mLH&} 62{!<͓ Dێlգu>Ad D\gBI[-ЯAe*Y #CE(>nj©O>hT}zop3nsջЄ1/tOF gT"-89wlœmrpˬmZ0vee蓱-N̞#s[$ pG&5ta 4Pb/|KI_,O޲'.5"}=%a'7%Lmҭ )-t*wnX~);G<޵A!_O`ZVSMkPdeղb*MuW667:J6eۋVeJ1M?, M9܌G`,C|yk9!ˋןWJVlEUeplta[WA < rjM'oB5Pe̴Fqٳw8B`hH ;朜Ճ"h! <ۋl3o; 5dXXTT囐-f: B9PI P?]+/cG=G-'k{!4ۼ JPJ =Lg+ZbSREqN pWM6S#2X-:dD-V8j#܊/5BVtq ʍV m8FKyK' 4t/n -to p/ :<805~!kzyj'5_ңpg}:sV(%@ۗv`k YzVkE{mFwU cI^ϫcX;R*%ӗ|_M ]qU7v{ҊFKw|B/b AHc%]MqF Ŭ>^l5U eۢ\f{S!&kBaAaʮ[uN\ddbH\kJr#ìEk E5'tx,ZnrG҈vn &VCSLaNֽdn0eH@v7zm-m7Yhj%x-כѦ|R)k!4/Yа&? BOM\k UFgښq_Iu6  ?־^-ŲXJptz=bKSIMA *s◰`cG7N?t&oX?1F`z'荹A'M*gk.LuP1YV']51##AR*5}!VZ䨏GoA0&L Z% 8ی\T[*`D z8 (؊͆fYU顃"(< ɕ,G51Y=1Wh+1C%<x;O V ɢmF>~{ V񡶷g5oٯ> ~ߗBZdwbU0L U14^}!Ή[$FD?~'1jلc FR%C "?cv'z'C\T@fѠ8hMh; `Ao pN -Р, ZeyL'J~n]y9L y?s6 G_kr~P\ O$C|*XS'ߦyƬqGyy9RLp;BN)b#xIbc.vȋa2Ep\mqbaN?5xtí;xfQBTT\S : qs"{u0k4lT @ Ex}E8S72iuTWkHtmѠ<*Kڎ'~4aֿ1݃PO@tjl7C^ ƈe)`2l'$N.`63GG3ͼkTz6& Sor< ݼ6W71^L*ݳ;~פ^] yՌd믾S =cvVN'H\[3 IJvcfabSpX BYĸFȉ9 V0gHޞ𾕃tiV-󰃃n"57@QE)掐s@/ࠨX2iq:R['>>q͍Ɇ[ۑw]y y9SyGvRD it/6ѕ2u#6ʸa2Ɖp=GGWѕ1!chZrѽˎX*NC}3zs0Ѵ;vEY)"uqpSB?W}{EǘtBE[.<#sV˚,Ze2Nb݄U3#+z*TjsWc &^֓s[МWFgR3W5b״9sUgyT)y間vZzyph> n۳%m"`2amZIJLUiZb11v[s2k̒dF:.$5oB c!v/#,{zy·GaiHIB.{ b#6wwK\h˒TF eX#1k۠;v>Jy' ሷx{чYg*^ 1H$ÿZYͪhVmFjެz^ybh v}=$猨gB)֤D-TS[`k cj/nx\TJ aߣ1muiNvWZ҄e2oO~7Hl߯RgY||ǘ秗/GoNCzzgJeLT?4roOVFxj seƈeӯTݛ[T=F24Ԟ1%>}9ke9PIbUZ7{ aA.\[?]utxM<>״ۧt<\xҾRL-LU$Lܿ͗JLEq)`l"gpý?ȗxߝ< -$-s΄Tb5'/>n6.=Mmdzv=Fs,NY*К䎉8ZިVc5: وKQ.I Joq{;z$qw3XtSF+VT밖8J1TrU)(=$rKҘ+ ?* o8nzUmSlҬ41qұ hֲOu#?~g{9F~/![>R{磀#z.l}~JBSOɦ?Z)hZ @أCcLid)Ox[dJӷC^ P`q*uJUG91hT[` ::`B=YĠk,mX zR 0kQp$lks`QAXb"ZJs8+S_r`!ANK3%HX>dTik1*bfս΋Oj~4] UEIJL=sFjej\RJ%d$\1`k!K5 q3̡=C^XUr$6̳ijb I ]O3R٨ۏΓ&Zc*(pM䆎,3Ywk]=; Cy[yCC9ϊ޺㖎sh_P>_$v`>UlT^SBjz֢R()Ju 6'ؤ-QOE ?6u"+]-(@K|O,g=]{.k/8Z,T&ONF GeQU9e[܋RDW$Q\%&Y^sg|5%QtFK":Gi ++bJ. ؾw?\c0`??޾u ƾ CogAƚ[a8nŸݻdݱ=yxoo0>Y>V[?vh8dRq :14K-#v\mv%ׅAosܰ?! {Ln Y)AOrP zM1cE6w-ULc1;J=,AmCT5-jV7EIc0hX#(h+V)iZ8{*/esC6L LUJo7tq_'%]U)WkMHS]M4[ X8Hk]u2(Bb)#7/IEBIA: %@D[d"aϪ 7t^_O+Y_pѿo㑫A΂q&o ,umǢŃ%C^W( l/j̪IL(G*sDaFK2'gߊ9;Ϫ[Pԯ:(y F-ڛ~EHHi;y tMGȉuF[o9%4ZB I#)_~syL0pW姸Hkgn0J0Kl51X:^UpZrh%ǵZ,Ee =sOCpǘGye+K**AxK [T6(}FODc~Mr*>U$;}۩Z+Ֆ$m!e 3Al6FhYKF,5 Zj`}ĝ[B1DСts~@4[hĹd6BMm4+ wY Ф&0;h3>X0у_t`;y>~~[CjMdSTFA atvwDU"; \8*,"2 ^ĵUl~HcIQҊwaGtt Ύl Yb"(G?a\tҾmxUO]d~<`)<&q:D1P{VةcEN"zLxE){7C|6<;k?iHzQ3L숡 (,0ǃɈ`EbXPYgRv+-}yWb}ͦZ,D%cWb4 ,1J XOu7}h-P{'p΂|S'k:ݕgjz^dkQoQzcN[9V3pv,[  ݵHa'?~5jzo\ܖ`m.Z0  {L.&ȹyϲ9p `ϟ~2q{N;o7#>:47ŐGuD0ŗs;B|Dsc&Bmž}őBVsd.1$KѰ]҂X\>$_Bfy f+Gq"@:))1-MQI~!54)'+VތZ n [KCՖ䞍DfF-vȳ&QG@QdAȑg+>$`RoOTR7lSe:Zn̵LFrƌSNkB7Dfh8Gm$Dh7Xe9Z)dtwcʨyW;'kIp|v!v=Ps R9DŕB|@`3vWllC]`7ۆŐgcD+ŌҲw{ۆ;Bɚ6AL5H $ $hB ZHRilA] -Z8@k|ݵc Jb +\ieHsWX^kN5u1Ȯ$Α6V 球xCVLXIWx;y&mq |WiPAP,9L2`bէ(TqzH-<{qxV 9L뭲<آ{7TAeDfۣ E UڬSsX%u2D"nn~s4iiB;7%@;?SlfF5QeYۢÆZ^F&~G>ljEral;]x1;{liԼ慈؛!\ ;-3gYNj.сeMŎA={-tNo8{.KWstV>?޼nRJQWy6 . OGps[h~Gm>c#9)ʊE[kZҗHP}խN;䮟VrO*x5JE;`KG=G[hI]݃5WtA ( -%{7М' %9um#9uͩ۬D@ΨynRS[IHh`M֌߳Ց@ՆGUoֶG_j;r풡muɰ%YnP";W:w$MXs@ dmȃs}Ջ mSL0K8 Osnagyp1,|=I{1 Q\y] D6,t6fqAj_KMy=|`-{8?h}ߋVqu㔻J_Jq}QM z_ED_z1Rz!9HFi dYԎHJB,z_,ELJŁP;RxL;WKݾMjɩ28ΨcykshR)٩Ugȫ>@>A]АQN3!:Y$T6I"Ze=&iC" R뗋xBKŰ)hL""lT1mڅ!f&i+ R;/QAL($I$2G匄7%Bss8 |rDղUB61/[}! ֪B(TкYgRj2i"kOlPŢqWh¢=hNU!]mA7rIdcTƉ8a?PC׋5|Al&> "Rp؉6]L:SDҕa)9+틫}ZclI8n8^C}/ʬ̢׳Sy<8q >巫@k1u'WL)zgF#$[p-%<q@^IGe)8t. eC}~= L~DRΝWћ^V*-.J]·$K,.D9}4 e!@.'TB ^@aI:6siDϑdaP Bw%(s"'{{Z>msіE[8 ja?QSQ'LEB 6zH\&7~"χ(Ph8+VJL A+E$`h@:`̤o~KyKjQ+8bCmmsE8NxO_ˑ7zk It8A `{ЋKDCȲa9gD^xP}>bqH"؜B k}뽳/qS] ˍ~ɕ81Nsu<gSnw<성᣻a0Yt@ D 5I{{0k/ X%Wq|~2K˙KWu aU:pyIpF B_HakZD0#7"XQuPr邼)y#Y!PÕ+`u*)&V5@XUOI#`2P :[gK9x|'RI/5b' o6tcY1`^_] UcmObQl7Y+i&XL3Ne{yN`(e6GdXtKZLK ]ϬKq!kzxДv QجwT掎Os"!l^r'd tXќDlV}ÓfMJ,T-)6vXW7 U׵w-)ZP~/<{N;Iz$xAӚ1{^B؈ֿ$ɊkƑ.Zem^ߋQ},MOE.ǬQ޶k#7nY9PiF:6;`לԍ7"buL CZ4sxvVrk?_Ljd G .;|eoOW}c.qso.(݆賂G Rxޅ4sz'}dO(;5٨n4}c&V;.ctʾW.X^c&%jJOPI}IJ,+ǞL;TjJ[TU0m7k^N7͙RrT!z0|ӍL hnaUru #3xuY'ݿ?Ⳍn qp6nă3:>n͆of{QtF>- zpn-}6}\srx\6v.rdz2Wǐ~8oe\]_ Tfi7=yB_]pY[g_lU04j ֳ?n?&_"y{XÞ}VޅxL.*0ȏP\_r_od}KA{8ق4;_/,e&6_?2=ݾz?>5|њX`.\a I4u i.&zlX]{o6*w.Iwhwi mPqbvf:SwCɎ؎%a'0h;<<7OhUf$y![ԯ3m3Z|0NJt^?1Dqx[z"1GEzeLĄ6ZIo2Z+4#V8ccOqf+05'gncpRNwEZnu|냛7kPT!z@8[T A0 K O9 :lŽf̓(LcwG硢Tg m}׮,q^t" it˻G<'cCǽ޿FR:nt7}]jTޅ:?f"K_IxO=Ng#U8a]j&=\ކߗ1` ȇksq8@Gu0㞇hK([MY Oi%nPǩnOhp>2q0P<|D/_`0Lƌ5Jmup FOY(L=DLGUtj PEaVj1b2%tdr Q%zŇ?2Tb/Y'0XU.W+mZE٧[h8^GBQSu:8ҊPsqlBӆZ.^ [}HHG{U$0\/v}h٩;ga6e"c0CsYa1 l J`v_0بr*UW<} *[(ȨX{fEjW΢!FK[Jw3Xb8c׻*8}z<2IH*̴+le,u>7s ply`wU`Q4&Y&fn<9h \5 qZ'Huu9Uh?|c_t݇wLNМVa8Vi3mnEv( ! 3IkMMa㻡]I?qvEꉞ\J8N=rYAWp㥦@G2e[}i2eMLSWUk4k+M-!mKt3*ePHMPhV}RCsCW߿h*:pj$(jWpqM׋qBk4W p +!\ +,0-貙$/Z8~'˛{n6s#@N' qĕ(J0Gđ͆Ȫ"3G" h]m1{crUbp׽`d=Fϳ?Bo/I{>j!7W)◫QĦat5,5YĮz@B.zFC/Q2uaNI>ҁ})(N3Ґԁ'(G&V:PCᮭs3@Q+Y'gU/"tdalCݸ7ҟͿq1W} _:ѠŪ+z(?o~8,~GaJ4OH$І19 O`(obٙB΍+\4j@Jp#C[. }B Vn F¨}:mitNku74Rh[wlkV3+ϻa]CYH러IH6l;ƃ.r4wPH޼>zTcHLp~`xBF<]ƾ5Vͽs?EQj&fJJEEu]d:]/T0|eeŒ[V6}baG@M'G- C/,=l9GF*먰8ӕg_/j;y/YTS&c%NYbJaE;AKR(Uk gŏӻ 8{lp0xv;AZ>::+ް'ZBz_'fók<& WJDkD̉G3qg؇kBz‰6+J#Fr:_{7ILwu Ky-?5 .aP 'n8Csn WF L84ϭ ywU٢=YP&qNs0Uo>J'AgB_# W[fNG>q}q Mܺ*~5mWή@~ͦ KȺ~V.:[ m̲,1!APAw ȒI:{|LOa0O,L;B0uGYvC["0&\dz0H`$U":z%ܭda}"*Aa!dv跱f˶|2 6Ar[2sgjhj-jcTo> )l"#&'Gz"9‘ED{ cRoዌhZ`iPL3;]v5Ͼ ?j/"Eڃ(`{nlIpr}&xuulz }^ @ `CO5J&heD4Eܝ^rBtGUQ au2v{<̈q@!3L1  vgu9L(5&mZ 4mpN`. ?cTJ ِ•i޸ h 6on//sr^Ci \b k)|&;#oq8p)!D{3'M;0_zhFLL% j/.EA[X ANeưYFi?_Nz'! ]Wi?|?)T 'd|7yoA'Qsn$B;g>XĎj< Uq,ch$"9|IOC I+8W:!?S<[A!7l͏:uq FDWƜV况o`,n]GлĹDQ 9"#By г`NXtN B-fT1)bKŝW9jVz b~ ADD DMEbBP+$!2qZG d5 Gp̅n)DHoI95Y}Qa %aiixn}lb;1|JPI/Ng nӫόD [>8gj}2炎kީ0+CKbV}jMNYeIIkypF,/K%Vq4`΋_Qt u}/BEH'`_&$o(Xčr$ky’e3+Jh簚,o6BU(9M{ C*$˔83KLiW|wn֔J%-#1|D~l@CN Q, }JVJDEFSU1uh!:|Imqda}젭 *P`Ѵ Π_T@lzOyI(MB #7UdPlJ 4-?-X*ZPJ zRv1, }â뱸u0P)$6ߟJU(vw2O*, \e(Ui4Y~F,eK?R6m EnlŒ[wRp,UӹyN"2<*x :ShY7:Ô!x۬ZÄ1hK3"Rم-b5praWJsd|*F> QH*Y4+%WnyN,HcJz) ʧз'a!6"K2Uhv f|-5.~'b#p^)i@ ^akOw6DHQߗ[](ln<*sRmx*A &Dj%+%`Uڅ>E]څjBivp޵ ݅vZZ5yr:PZq PU5Cc` F0 2`oЌ*ݗ@V~A B21*ӧ5m%fM)պa6XlmwPʄnjQIb!8%Jo^܄֒+YRkoA܀6{˔*̃?7t~~Ô".]pvw.BEȇ%8IxlcP2aI$tGNFS]l5MDVs~MP48䯜QEɅ)MYy1nȥe MO1}Z Pc{siNpzorYIm PHƺ닫zp}h8?Qr-3ܾҋڂ7IP*@a*h Ġ+5HItڳV_HYmjcnrlPj<-ټԜjQ&EA F]5ԹKK@JQP:)|c/X4KN#.“]u-3K.T3MjdI X! Ku$C.蔁x42q!26uT*6hukFT.c.~G%e&1%k$=cPi%7hJn22FTXw)x'C^2cq3&e_IU=eh6rNb06FEAE ZNqiRJ)Gpe;rҴuZEeKN6]&^xRbx5Q ʍA#. F>4li-J%{i6W$,gɕ^RROO.:uA<;pC17r' ϨMBF;,rUKw|87|i򆱑'o2RRGفkYRrbIk0 \|g]^lkˏ]QX6=eGv_d2uLڥv7oכ"h;pZDU];'%_pxGo!Gkغ.$%6.yA6mx2Dm9}rL 0\s{HͫߢP 2<`{Кoa!C7Ti-E H3L2Ϳa;1~ htoy }a27@# 1 b wzYGIdV<*V]~h0%5p/J?T3Fk51Ɗ]Ǧj"\)0;-JLlPz">h6d6rY=<}6+1hy f`B,*[  *Z6+F'\nޣ cwb v<vjQsWtakcG<透*dUKg&+XpH/,'eN:GE/֧2W"F>Jv9@My%|VT<mw D5ms8 Rk} 4dCf0!ɬ(f̸AqAfE(akD=G>5*5pȣ`MRXaj,O =R: 97'yq68㰦\YMR .C Dޒ>}a4"&4zUf,gni0<Ǣ&Ua$OC+7ar'1|ț 0?qXBP<t1EQ: S \dF@ܞ:౔ U WB&N(Qha7ҥD%,h(bΤ,A4h^FWU]d<긔s"]wTg9$)\2E lp }lUTwl畷(AQTxL;*sRSv$O.lvd1 _>G0A<^6w) W--.3@(!o0LaIBIRudRs5ב{ c:(Nv6+,Y#ʥơK("3)iJ%W!εPh(GG핒e*i[q2ǚ8DWZSM3Mj$Q$1ZGSRbcVw&:N8]h=9/%EƜFM'Jr&]鮶DiMWGF9vqwC;eYUhd厩v8x CdK&",9%n7Wo&d$*hvӐ[]% ~%}oDH?Cp}S;j ߠgB FNɹHC$bb [ r pČ`ROv[^CmOvP"@%4Q" i`baWZν;?~=8!?6;q3܊# <Bû gz~MbRti:/,8QYyH'@K/C+ː]-D.~a>n7n~;z)@@LlNh-)d9x%:޽c\S"&GLO_/V #Gho z9ʋ<=GsqHp}Ď$ӰSwvp>,vޖ}AeVV/ҷ`I鯺i$c9Hѥur"l$'@dRm_06_ <s8~XV wڝj}a2 mWx5c蹌'wU{wZ:) <^Hy<<`\C)b:y jy@ c ԊBlz1 hM^:uRi 7R.f)0kZj{,5/m-1T_ Ǽmt]89|V_Wh7v\P\DxUJMZWͥbٞ3o+ ۶Fo1 Q/n{}?9}4"[Z(Sj(S¤g5pց )1*3-@1" L`QmQS)`eCj(dfϤRX qFe4j`.`Ɋ3**&؂%=;4inw@ `Q*@~@9d=,U6(DK._n-РZHņKD5=5Oy)E+g3wzA R6Sl ,+"y7ԲXE|I ŧ^ 6[ hͳkRjh OJ`s(|:]Fsil#ȼll"F&i1J xMyK7 /gꗍۋl>Һ1iXx̹\;JТ1,g^κnmS›$S.tYJwUk%qKYj;y{Ԓ߱wwE(\YP-3ᑡRw*Vi 98lU3f1JsèRH|?{,~[arrĸ)x%tE2L)k?J6#0UGRZr/R-]'ϔ[\!|щn7֮SުnG6&tFYT~dW#e=* @,2JmO--WJ ysv x~},Шcmz%`ּ|kFuuD1Y)O{0VKgznGf vcۖ}@k@>i)e= zKco ,Λn˭QZ+)6H+|bLs 6pxIUBR`H\DLxۻZZ\"XYW!"2#PMzQsЕsGL;RI06ZTb/PRQf~9+C_Ò|wx|.% }gsiغ CTc$9t󇓓T*Wv-,uV), !uf^w̎hTM)|h@x_ZQUr) Kji^p;Zq5diԶrH`=}vW?jCڱKӿJ|7b({#.RUyz/ݱr ˎ{IuyGˡ_ˢ}iYl> ) ><>7~]_7 8\] _Êh*sr:: T$*ݚo.>fos+j6sQeYa l6 a+$TNb.#P {V l诒+ ӫ/twkr2|UՑB3F.`΃`gFhs2mnVEmOG@Cssgr\L7$mL$hDTRvII,٠D4U>$} >qu7\J_& _m#Tzc6!Nr}myph$A*a})MqbR!!X*/yvnᎄՓAsa6fME{xs_̽\ެ̿܊/W ]>ufR𵾄>yPۤLy[jyS᧎gT^S;fZ֏[G]v4r^-|F1!:!Eg-m_o!(Gґa0ogµ)Psyblbh9xwlJܫ+u=7dp*Ksf>Vr*#uͽb s~ǔd+ߟIib|?:])׌ 16#"Δ2+, Lp,(JD !cxk)EzhSqJ)R1a+D?OgϞKZXp% S#^3NL y O'xጺu˜ӂjunM{0!a6I켷706KYXD$RTabCa%Z()IQĿiAP!oNH}`EWWK\QR=`+rS3 "KD X{ B"##[9KZFq2LR Jh?&ԏFqP8  5XL%lI(C~݃)~" $Ù/IS ~N̊S=qL}8[Tll&? |Jb[$=$8,jت'Żv^H;ukm800IϺggM#Z*%6ׁyTYf/'ftXaա} 6(4G7H@gW-/l ֥fi.3o!ׇ$NsatHX&|pbA4}Zw:Z7^{y]G|NY*0So>r>x>S7;ShϬ7zgIRs d:NƳ^6??V:y? ?w*E/8͠7F{/vd)FMU~ eX$ɯgw#;ޤYrk5"Vڕ?"p>:5@G;*j~^?޶'n^)߀(?|wyh}Sn0KU T\ Z1AVe* ѫ>/dξY˙ZerxoFb16y2*^MY4+pfXVnev]-d`;+vKһ.l.j.0z*8$8Mf&p(WJJ@w򣛔$ApbiT4diEyBQg5Z}[ M#Idredn-x.ܠ(uG]97$ENd[cjسmϱprw]izec\8f^hl yvW.W.UXV|C*n:1kQjC;SA]Q8I:Z:ωGRccJ-`w'g ݥ_cB' PL 0$h0ڥw=1k u r}Po+w#jGo];QrMʭ'\wbenY2ۖLȩFg4|FRHZn%h G-ZˋmϻjEQzKZ͕ʩajEF=4tⴿ.%»K2kBŲεf4\JR;DatF(NGT D6nͯDq) g`wj;Wۚ]Vk#z Bxk鸏~vT{3u|=8lU>mY )t9D"(ke\V2T7qMlkpq$ƒ()DbJU(ub&~R1߄(b#q+ġRŏRa*("8i6&MS &9)wL`DĨs]ԁVm"[0Ѡ]Ng~, !)B ;fo&*91f\#4Y#5xl9jD#j@%ā' (&TcTr#PBġB & 6:RG)c/A-G>t8!QYnJs~y-7f2D;apZ7xւ@F 0јW`E؇nm4`6UEXC+:K7 !g͡Wb|&묕ZW8/0;xS\Ҥ %'cZ59W5m"p{v\fD[pԚ/ћ$U&ɬdzLf~ KY4T Rci,#-"h]Qȥ_v84kcA$LVIfþFσPG'1mλ~R7__q[ܚF}mzQkk;zOW'I-.&._uoB[h5U~%`Oۛ~kRԝCȤ鋓tg[ȴ^N 4P,{`}{[g$6ٸOܽ(X J/?xRk].S۞Y\uE{3H3d- ywMrm[>v/!Q2dTͪ7\ H09{R>?%d]?_Y,[D^C֚=_tƙh[˅W|{|=ކ>💴~;LОS_g[x S}gm7zgIRcu2={k[y~.|4*)"{gYz3oQk:|8 ׳,5"s+ZPu*<gqS~ϬUfQ֛ SieEK)W;)Fڞx4.w٣נ_}0u2kA,xgyw^CIx{z貧b}TtP J:O@o_*ަ-KN7woFYAe+poe;ǽs4&^7 ee?<, ut_8/̡~4UDA9UT1Li1 Rp~ƁÇW^)zweK+x@> _/7\MR\쏛QUbU}^h,fY.eq B<ֿ\, "\Ol40 bWX1JѳCT{``|1pA0 #I Ka^~~V)KIy~:Elk, s2Z.m'V E^6FHZm9ڒO r%і\ÖYmɣ-y%kْ4+lI?PƏ "b086D7fK *9_Ӷd.<@-yYF9b:b$Ooc/aFVY|mҾL#{?cCj EbaYPusj*1#0lg^x8ltrr317, G30:蛕?xp;Wjv_]Ti2€ ILk)Cl"/>1#Ƹj#2q HqY f'lƛw1&/d]PK3}E*vÈ5X~Fhvքw[-l.GpPkK Ÿw1Z Y]*a8V?9^{k%fe*\ԡeʔ 4<8‰VZ,h h<<ÔFT[/λgyf R C ִ!4l:wS` U 4 V%/bBkw[JmGx7eI4PۇG6 1,@b[_䰅D ]P>v.à@ؒZGBU0DTEIj! ͦň{t)JqJ 

wMJj|pb!{Īʋ,2 P3hfq a:gO|6#TI eyolk*ǸPBz25mEM-)Ֆc\hab>HUA\blbi AK&wʖ[O4kE].ς2xIAi89CCoZ0nZkW{c QsK}=ۼ*JJ%\ ^*`7T&FT)"0I'6'ه3O {`pS?Z?3^kPRF DYR(WRqs'3O8O3yh0hswyv jw;pD7MuB3vhpÓ+ATm#nRrMGrQHv]#!Q1يxfQ|W5 D 8)8d M3]G/8O1͸YHrރ4Jt *OT@p-lRC{s;xDk}7 $@gѠ< CA9aQΣJ#8R2>c;xDKwKZS 3Rw.:Pc䏳bֻY;` s#+JL8|1ou#|Y;9#zj_j/gxg@lR=+ht9QJvv{F gRzS"&bCB Cbs$XaЄ1-Db<8N^RP%ؖmNHgW_\,bۻ4e)pWzwoiWC@eBƟosY|q䩳k흼~il.@ !D(wqԜ7ⶼqQBFtCbm՝IJ.R=xPR)3fYjH Wf)Tp'QsF)0;0A%f ݀PNJ.H 蜎4 NU9U{ޱ5:R8JM'տ&_wX5{mCdm1 ;?gǾmg^>^=+/5GyONQ ٘x9ԊXtJ 42HN5IeoH4ذrtE.}S&Zng '|(<Hap O2\#߸z%笲*VƹcYb %%2PҊqK7EYts_A^ĊdG'*G20d mM DUP5@uTO>Bsyf jƆ!-,񴨨 7ZZsm6H?.LqRb%7G\ XxSx3;n ^qsevok_ ad3L/̈́CMtHF!@w V*F"mWT $SIc.H0AG,#飤@fE HH [5QJ9Pz8>~h)uajSzmeCyp('xinuocflEN$`G<4ChYuka˧) W*:./R׽Mz|Ҫjף܏"!)]c8%-_":D=tR?;#c!yS9"V=fvS{ ͅ?_3tQ.x:QCvS?D^wk.W'òIjaeORH}Jp^D$b$C['dMJzx_՟Jz8!ALGDz]i|Knk!a4HxMG# hrjԪOԼ(D2R C шHl:6Y'<ӓp6GQǦ'I{JrퟢYFgR"<=9wo"H`c^"8\yhN*&Lv' zfS1 AQ$ vc"v $ ҅CpkQQui^s rUx3ɤΣ!'֗ayz&ƯK{eg`?0=sG#{" ,,⪩[9qCk]t*r4lRA4,HiCK֤! OIC6zBzF9=AtE81 6hibT8Kǹ ,/׆X$M3fR>j$DA@KT]8D{F(&J(!ͨ` a-R VQ-fr\gdN=tCFbآqɢ WܠSAJ }E֔(.pKNRڵܗ1A>q0dq@ab0Ypvr ^a8'D[)e^6vEQ' =qL9L/ w{,cY̚F5b~(gz I{LJz9TgN͊^{8~8c2?m2?Od,_G뗜5ai!,ZTsu70P!M?q74yeZ[IF5ۘne| N@#JJ;>!olۇ@ AQ2e.fKsh#;i?WzXGR(-G0ϒ3ypz~7w ;$@P922u*(׎y¥.Rxm> x?>S&馊7Exi\qCz㫇/an2RG JQm#ۿ"ˢإ<3/&- 13Ȗki{-J"R$%%$Dg9s^sը9lމa^"v@R[S]&'i,BRfN'fgtIz>D #JtU{ݤ${SD@=acg6')w7[JHثh.瞵Y,-ll_[b xݶ!:rh!lz@T|G}:;;iVP3Cd@D,adcj<Ê,~_Z1uL~G]rϸ"b]7ANoX'U^+w?I{ҕ#mÙw&I1D1b'-V &$)X=Xtd[99uq C/F=S̹Gjz.Xr\Jj$1>Dp%ъep"9J<'Sdzu'hIXŌ#o`}'+16(Εc&/؀vM F''`U_t}t D ;?a=VADclel)DҺBON!NM\i1{x+Z9kCXeG*:y(b>0wߗ/_A' iYÇq2sH#7?ӹU:JBW\|.4M2Mr+|g66Lj)E~[]b+`cr6G-#pmKܰQsuP]I VslȒ}X{|P''DUN:q,Iɣ=#qDa GB5,,A F=X q#4 ^09hN$Y!@8Unh2Ж(V}A@If;^/^18*ƀ*JsZ]:@{yLR$(ӊ1$_[[T95vtZ'_)-ޢ'@0]R`L2-8C1L̓lWg׶t;jCRݜ6XT^˽/pR/JoM-g]!ngMWYX0z%/@0x%{-F~&3³g *Jgӏ_y0oW1giXVrQ-glߌnG3#[>X[A- ** gƊ{2<0;q9>kWQ[T/3K/lSԬ3P;tSZ➩S)gT*\T4,+j*]W8=n:֘TqO`w 7u@?UkM?8~^݋߽\kLG`Iz->BN)x{X!UHuËއ#"&,2UgHFž[_WaMv.&P&/׀H46gyf CWuVOoFL$Q68?7*<!iP \HLP@K ~un R1A\y{6[blg -!Koe=[SRsU:3wu($ӽ0^?4(%޽O xJsRL˿B7'fMGۻ釔Zͥ5jU܌ӎ'o ՏcxѬ.%lȝӄ:X`_LΟ?ߣv?{gY9fk^]pg=?߰⳿ fwvlǏxe jhBho\{b53V:NYߊHNU#ѻcj[mIsmYW$/bq"ѯ,C9ً+k,!Ġ˨.{#צ2C)ɂ I=E;Oogo<P.b q+!BGӯ {VeIr:\85_;Q j073Ȝ b)",|WI# A!/yYrn&3]ݡ6Z j!Qt0֢{㽈AH= !WD/2 OvC8thTl 4,ꇗ$j7 +`TO%pm7@*'WQ*'RO@\D2~;*Ԕ@#0DX <>?=j ,e5ugihArGkJ/qq?^̃tRS$$<oI0=&v&mlR1W5'W#+Tg{a=MH9;/E[T X5tJ8:mDB-Z/tCpmST tӥnmqPu>t;AMɋ[^h]|,3OIn*oOkV `LU|' (𵲙ٱCbd_Mn0.~O1t0/0uo<;g#w>y:9?{uŐ|;\T7!wL(΍.NsE*~ŧI;\~yfO8`3.d44#g2iP?t^`8{^O}<$}1X*6=tv!}h]V֤RIVՌikJ129,`wuL+Z^h "Furv ps28H$ߘBBnt~E72RE5ʍ `AFhQ@ėZGn@z:<ގʧ |Bm;k?[PR6lz<<,w0c!ʸ w+ly.9"gus#NA:O}@(+EFx܃"h,F^Q-mcܝb~"7\KxV-9Ygʐ^. (8kH⨝!RVLs%|L6Pd %@8glbՏUبHWЎ4V3l`Q4C[!"$T4WGW^ya׫PPAgZhcFbk!{cY1fW'L$ԯSgkj]:BUduKuNj0Sl@Q#-GžL pUhUhMlnmRxWի{ "GDr N9Hp)g}T!r $DoD*l֧6Ca|F dQݡ) <[ $Zi HbbRzSE7,r!XmM$u j gSBFHNqJ)tZ Vbs# 9r mA|ʲ@QH$n3(C1SᩧAE@9AAn|W-KXL\`5H3SlM XXqV>T3)SqV1TL֎!Im% &'d̠1LԎ&.Ԋ3**n.h؎R/p[qOňM&3& mmh*zSֹÆ|J9PjtYZ_>rB{^1cFD]AvDJ,4Sm֠9 &#B-W_c4WMP3)`H)d=\-n4ܣZ7YnojpkQߥWt.ߍIs x4ϩ}J=6L]Owh Z}>ە\%`W\<%>%:!qSV$fnldÚg5?!)*Q1ҔU@# ~c2*YnUN\Pk^wޒ'iЂgԨ#Œ]IHz%,$ڍ+ĢZngQzT+h@>Ksrtāq-/r!TE΅\!7h;!"" wEldP-%8Ӓ464y$R2)5s,%JBvp*Zuu G1_op 5˳@ ˤa~UuY1|NRJ` ؔ .C;噎u0mQ6C@]q.UhsPEL)}(74*-#BU.!3P\sk0uUX+B}-=vV1xE"@QsrocP>M}G+kn#GْpC힘cKw(pTISvO/PEKDòD L 3H Fhr &&jY@~S/4D<]  kڄ8OxHA 9>HrY丞#|!aHe3Y i[aA; inÐE>8j9(czmj4 rlAN{{̑BHf-_3D@2j0B,+D0P3P(2j1bE(#dr *!kYd`-hku=~1Q%>qi1RC`}n0"wF:1jHð#ZJVj;rb5VFi i%/$ DМ"hkCeN6AĨ5E5fgԬV ı<)G=jm) ko\DdJ[ڍHqD,BD'&ڭG1δ[4-|"#S[ڍծvH]=Gq"qb '|"zLq ǩamsKj0X6Ijs +{iM'NVeZG 6>(0 Zq1SdRI Lf `SUw{8_v 5%Nj2(`b9ִ9{sesP8Ib glz1Ya)gf;/s3a|+I{+yL-˗Rtw >rBh;Tz8foP7 U׏Ϭ(iV=hk hCU}=NU݃$%X?~Rp>Aպs#U>ADOS(n4ì;-`]Ygm:O'N$!yL!9NαA/:Wih6dSmH;ڴjl7ڴ^7vo|o\~XJ^gC\k֙gLaM`@:sJ>}8'HĒ{5ca̡@ᓌ[*j0JkB)B &X@(,$/,43\(Jo6dj-izb Ic'yFG>jO<a50LB\P Rv3!@֐X'D[ sr#PL(&TN|))cLԄF'2z xH'ׄALV9Ks4"J3 lB<@&>l1CސxPb =F|≗ m_N4i)Z~p{dҾ=$`u}3j)~Pe3FBhoR5?-ڋU4 yC~y߬/ jTѫw?ϝi4no =BRB(HjitLHjUAu@ݯ!(1D쁥(ݗJ9S ؇J9S($))H$zBkw^%Ds $"T g"JCgh3ܲD9{`aI-TeXp8L-o8|O8|aFXYzPb)VV cHR_瓖^FĮ-áE;騘>q”i SIQ8uEFS3L hs=2k'T'7lo.G%0vƛ-rθVlI,DɖdjpN<qN="QwoLV0x!YLrBݠ'!o2GuW'jjuvQEͦSznf;6u;ņѲ{:ꄎeB̨`;K̻ %CwD1dI,~_6FnG̐\+5#^?u6WWbCO_ʓH\#&+$Rg"`J+tjתktUVhi v`0|;ɣH C hkriER 8` "72Q$wJ]pA=]݁wEտ$ x?$' G2h%<Ƿ0"5h~퀈eY-ӧJ H k|A ={φ.Bz*ۄ`{6!۲>Ql &|`s.OgJҖz>I1=~Y++9/~_@|4 =P̾D0=L "@z6!L؞=-X<1,{)?2ϻrGWj|zrH2\ϧm6 )ȣm@Bԅ$4>Xה{kjI25T(IgZeUb/eSO|26Lz"s?>(Kl/l/043) ,V|JZI;K+0Sѯ%&/7b|k7jb;z19d̏ 9Ehq:4^\6[8+ͮuYշ?,Q~Kԥ_.KT]q#7ZQfR˕2(i ^bB,lv%鱝oI ttx vJƽNny=XZmXBm2ŝ#神c^^Mrn4]ˆ#v04N-_]|TX_,yŲmy \$IQe~qwb_S*\~`PI[|9BcFfx[s@/k$ IrP#1@fvDb6=U} zԍ_v5h X{ގ>E7^{Y+Ifj2=Se!sȨ@ĺDLM FA@X oQNAb h+3*?dOq2!0nyM:]1nrQyA3n|ޢo2h ޝ;IzȬZVN_CFA"vw8½lv,j|({W Hn/d/ AwtШzX8X׻PMp=5k+mrDA=fch(NV=qrMdZ¬\I0j]&VSib-eGKVyDxsb JucE-G~tkE1]X~83p/ ow)بgq-o xwmI_aV`.e;&kW&Z$eqHI8!Q =|/u(xJP! mon,ZΫ@B&;K%@DR5/G\yfy_C|z̷㫢ցHIo 佫g~QhqPbXE ]WPt)9'a$zX IR2 +'&ARaדOeT(gPy xVI%MǴû:~aD&R:ϙoR~=͝) 6~ ~^-:zh_רqvpB}yZ8?4-2'DQͷ8$D/}!iB^!pѮ]#7OFS^ool*bt3pU} zib(dV˰0fw8w0t~ӊǠ&IY{H [æElJȭđ\ӝ[XmD];JUro:QJ!:15":_ TG "-LiDaN5ӄ؟NpI ֠66pE9tk7GpHUDri0h )Qyeu`AX+DI9Za VؖjխX cWa$)UQ9i] ɜ_o?%IBez5FYDErę$0@+eq2k4҆h'oatjA!!/9ƞԌu*F{_)DUK.tKTg܊֩ )NP)A%tJjT,4uJiJQȃQr_L7V +Mk~]y͓>AD ,8_{a4,'}BK3}h*Z; |;ƶ@' p@ݻz7JѹЂ11>}=yQnD|>'a%Yiw/g[0I:tߏ߮usYGMv>47ȍ K f?fy B{ц`OH 9ff=dXq8OH$64X}J_; h9 J;X%Jl&1J)O( (A:\*B탁?˜rˏj"”nvZn(K݀q 4'f 58zfضlg[HK,3FQqǘ0lN M$rpZ%g\aP|IIה+ⶵtW-YT4{D0jKy=y,,Z+FI7vcUkEs:fi aԲ?DiȻUaLN."@5p׸2"VFiR,dm}[ٖU-xt&=IdT0Ԁw-,E -H 10lyU<ے =/֕} pmT IDGDXS*_/I:Qȱ \B/͎H林u)#fBNu3FN6eE"K~pu=ͮi@rnU}G Բcx;-K9/fsIdyQ<߻N=N{7NM5t/+OlePe*)YjEP^nxUkw&&]z(/PL Low4N0g~s(Dpzc^l^. =72Qu9W3ZܮHn{;Xր[zwȩ'CŮp0Go⣒V/Tef シ};Bq7#+㴔ƍآ]tpvt"S\]^/,/)a$i )ZGk_lNwj^HE [Gq0=W"J([vgcpxKIx~8|h8IEzܠA("̍6cF՘9J+}4Ulݵ*_*oT|-S5KT]L16kk1BD5wl ?>]ibMj|[W<ՆR-S շXJZO@pUgwEb xW5l|J@78ddǦVW-"va»6C1B绿yQyw~u`nF=Z2DKi}Zi큗<`a!{E[Mf,4y U1p-H~[#h-\Ã/>9ߘ޶if=ͮ_/}1D#(``)H],b#[Ӽ7z}z-?u}?F0 ;]_:"Irںv:@qAp%ک`:8(|};ߓObȧUDGo9F\cHV wMB(/,\\у>醓OO>Od{@SCzSZ`l4`O5GwI IվDlf1Fdn=dQvi/ x8mA4yޒ3#ϟ:u>t1QvJI)TG2RQ%JiJ:J&Tt:61 9@x7O隂#/WHR/AyCMzPVjO:.\r93_L{lzՍӵjK1Ԝ LAc Q"q,%ES c:<0QI" :HZBL˯ۆ|J%rއٟl1)ezM'y3ޛy0v6]eypW7e 7[ ^Vf$b]|cKn4(#:ϨbXY`DZ&$ ◟Z7^8-eD'Uk)DǾmhYք|"$S ]|fǺI."3 A }Fv"pW-Ѳ֭ E4Hu3ZPgTng[O&틢|͆.QspW?&ίi:}rBOyz7 KlY39 g:E=癮P ~\0E6 G6 ]dtdSn?*,so Owi苰DʌٷnV1'1S}FQ܏bnBxC ;oC_َ~)/{ C6RnAJm*ޏ[.E+ؐxJυR)V-Lb5Y?&04eIM( YĥātI4 ckfF}8p*k; c錋b14*Brj-""aLlb v* h:R GåX53(K jJTwT(V YcjjjDiɡ(!I]SК+Q}JB1mo88g3˲> ƨ0= vO/bFG38kYdxD+)u>lgôo"qD fR0aǞ63D1L wɪhB'曠4AUc&ClmgoX5v0~d!|k}Z7H9)A 4W_y)屙0fZDG2voCJW=XiRTqMd,4Kn#w . #1KO+@o[fCD"T9ߴqPGI"R9$8uLHkc$Q }xW'8S [>vb̢<*Ǫ5v\vT(|K_d]J,c_EA(Ƭ/d!xCE%-w)YS$R;y"IBLi((1P;51a*Rj`iPl"Mk$2DGIP'AiqûJ䓜_7 |P' <<ўx\#"#5Xi"\i5cHBUx$D2*td:edŐrbS24!,I> oM>8#C9tCLA` Ť5PNBr@a䣅E>bPrYx9? j&9 BGL(FœXD&|zU3|Y?߼,{o^gX\x=::69Hi߇#\]ZLe RΡa᳕ rS4xmbS1ϏSᣈJa G!;#aߏ`45..^A3dl7zbFT !/UD_+I$ȄSJHK~ul=lE 0å^0&bQúxObeJ/BDg,}hc ߦ-I' \Z_&VAG"*B+ )eщ?-Vog%m#ȩP\2rv56{E>j% '<9>_NA|M14Kb4BDSuzT*FRBW`ߦdxw T\xn.0YTqYzfxl__V 8'G51ONjQa`3Ky0} l4L}#6$5s3J6725[6x 8nf/}}P#O `ϼ"1h1G.b- J:o uZ?ܫGC}|ٵ%9qag+XӒ jMs~~ә4hV]@fӑˊƳys8jzC+pkƿ( ƐnAi{AdMuTcJ6Q;pkI :Qw Ӊ>qY/?ii0ƚ ZXRwop}|a]b|0ì\dt܉ƍJb:Z '{M_ 8n'ޖݒ}}B09>'FIْ|:2ӫ&ί8܉ E ZAH_~IBƖ #COF㡹5j1$0| ^jYi*F_YkweJڭ>0/ǂdƗ2.>-ûTzc#0҅[JO= Y 0|DKSS@=iJKҿK_'p{J,dوvjR=WXH`'0KUr?ŐM!w8 fm@v)+YJ cՎyKj 0slX )swb.~ @A.B2)eK]5=BlGwjf5vKҾXqN)?Oi?|f.LQ)CBQvæbWZ[?ǓQ L!Zc/~uQۜBq }y ӓ}L#*@ $E=N땶`'pO}ZDEeFD< S'Y(Bp fO?pOԼ)}~23|PL}*NJ{6w pKV[]__|Ƀҍz[y1EZV KVJTt+޺Ґ.ZjG=Bn8L}c/ ǒ"rRU\-f/TjFyxžfC(EZOK FfF*adm*Jr4愉.u%L0=@Xb|8KMPX:#zKĨغx)l?Tp•e)6fYrTլV4[X1dBWUU?JFbV3NGSc0R}[ށ &׳l\Vwʕ"V!Fv|ǂCOV8Yb0fXhF9STXJp=YFEncy< <| cٺ,ibVSǴ$RjlJq0liVuUu]-f7p3oهdm6GG5.P5o 烠RdNDMY2)4Ÿyx 629LV AQ4)A`l+CpK,b؞U4G:6)ČDvܱ $G=s^EOl݆1/ 7*@XxOn Ҥ0!t8BHn TSYTWD `xBE"f.0\{T d4t|15*,労e7' qܚM*fp84kA.4 D?U@T3;KS%=y3\Awqf UQXq?gK+s5 `8N 3bvhz[H=eSKÃcU *h[P<ܡ(K &\2et>AfSDq~x=|R٧ӑ412|Ƴ4y. xfj){4y E`%>b:`Tz)Wҽ4V_jAn9&.Ͳ'fٜғ402=[Ck-Z.hB~-7҇❯Aφ|^vV "YԾ溜ߍM䎃 81&DKhHqHkDŽ̄4[U>WZu 0KlIG(nZZ)R޽O-Z(5pЈ+jQmgr%W!y5OIp9ttب<IPaV3O k8 6%y]$T;|s+׊GLI*[r PKpO\ Kp*L=9֑BF%)͋bJRN7N$">G?38X B0T@qq7G0j95Ogy"(&_*ot2[8be|K9ԥyؙ7k͏GǐB!arj g5!`UѳLhXѣTn.`"  5c;k!/SHH ƹl!НVXkTz] s#;xV "jQ?nNS/rA9"jrvK\b,E#A %3/ߚh^Nk@.\L7wy9څqL"g`23ȣ\ǜsyӪb81A3v/r𨩒5P" B}8hlN&C;ua<u_&ge3Ȑ;R"Is\dmdwvik0Xh^aЋhkcYj$ |㎈Tys7?*fT9εf.,>`y&"<̴vq;ԄYLVj`@)4XҪpܿ+ TV׃m3%#D%eRy b, Ƀh{iJU wgSýp0o)|.5#>-f ]]^YWP-$d% .sPtZrkKAl9-@JsKYGcTZ_n& bRssܤ0Ǩ u)l?1,ܞˬjF5DՉUkyڂIÄL.z`̔箓Π*J)p*ljLbJ{|0-D7yr:TZY1a$;fF1UO@  fJl|:4g)_%IEё Ir,~2]m4q:t3+sgI}e|ї:3ԭG 0 JCNpXt^L$n;x/Ǔh2^92R2N{ol}y`6ޏҥaGϷXcT7ߛQ8u;B`tmLP2P8!~kr쏮ӭy[ڞ,J[F*US>%CۖZS"LcZwנsrbCJ)::P ͕'cֈ = e}DfV}r`~:p۝B)x-xa]d8F4̉P~EL a,-ڨˌNsϛ/ur;J5hC87?±>[:_ںonuZ &A+R 5@iKtF y$h(Zn!B4w&םyWl0Nu:/E9 `$|pi!^ix$_^;bcmF7fTߘQ}ی#Q GL L3D:3)Ja(L@-zKlis.g44O;]˯$ ;M!] Ѧ a+V@`KSky^rwY> G5V!@[4[D^W'XVҡkz?vzm~BӜtV(\5הP泚ԻĔ` ׬(.1p/1EKL R̫.U͔P«jk dp#lQb,yn}h5LET%^oÆ/{a=\Kl7f0fUx| `ud*XzBLP:E+cZ-i\K`ɻ@*;)_JDpqJDD0@wSL/Gt5htˆ)O96)k$_SQZq|їæ{4i[cBkiIi12,D)MH0OIX"qe,S+;~${xtJB6ڜ+YIҥGe2S7᥷X Ap=h2+|7NKMD6E?>5L2FmO 3_Q%'| GTD"%)^r{Y[2J%7b+07q^W84uf`>R6 C ]wani^3<MkxyC8kP!" v`r}uIJI`M0"&1IH"Xl<9 ʫQHx0rqdE YSO!-%Ci!$aZ)TZ/<mKMqo<yEZ\yvR2=>ٸikGǼXբ_ȷdyR1^7цc-mJ!p[ɡ*G}ET3L{Q᫮y /3{M~t4yH7 aw$g|.ZqҮcRK?=3 ?ܫT\e{f@utet 2PFih@weIX`K@?,=fw {=~񬐕D7u,I1$%x*aD+#" +70%iS—{d<(]U;V2\n멀s!ʍZ2{ǡqlN*+׼jj{R'z}BUvS7@Woq3-9|9CĢRǚ$頁|r/az25|<0rĪI} м?H1擿NҜ˟'^}>a;3EO^+L/mYjv~iBЀ VUhCX5+TJl1]ΗY<HT쟧qM7g qm!E#D=E@o9[v_bUv8J} 6Yk #^*N D1B*yNdGg͗uL@+[o]t| z#k}=g\_qJɏ#Zy{JxaՑ4uyl˿T{C%)DxXmvI W՜nXu;qz_`$Gm6F_~bdv><~&.e(bC*ᝊ.譶Vj~R ᬢ=լO*U `S'b3_n.9~%?:]yrT U2')rZي#G` CDЩ ֲ@wDhCN8qa/TpfSmo.,{])f_|^R3[rxdۜ1ro[#Zj&_O݊nT:NRct?SWW!j83) nnj4ZJЀ$Ep9 ԱS+fՅj m B0aБZ;VǬ([YV+@O7AUH:œN4סD4BbJDXVAV85qRT(=kBډNjVZM޳YBR`NѐFC6ҡ@I `^0M39E=&Rk %+V^QiL]y@C9$Z`Ҕ^J SthО͜!ȩOJq` i42cB@ɞYR%D|, T d͑!!D0[l$)Vj^j ^Z K!:V2aʚ!5QՔc]H'5Ehvs"XZ);֭}yIU6`9f:$ov]>Y)Mooo:ͻn6Xm]-Cݍ_>7Xfq0hFsubaO,+ w[[g^z֦/+Pݷd!/Dl=N\j4j ]9m1iZkbh*1a!/DlJ4 ;KlsR wvNJ2+0Zv9B O U^ks. S[ɟ6q;x&_^O1/;4Y66;tr7d9'PՏcX-(p\A3:#M #83qq y&cSG&(-W)Fvn1-fw4ڰn96}yn'n:H16ۀI|TvkB^fTeei.R6,ZX0YmDɝ4 (Bd^\RPPs .%wͥEpÕXa#a-X  ڲI*A\ $KjDRcT1A Q 7#>b\kB^f$*Yd\Nnum T[vGS٭ y&cSG3rub:ma!gwoTvkB^fٔl}TWGB9A,AZF,[/2- ރ>Cj.Uۂ@ꂐ*x L [:-11n0?zœ 7уn(59CX*,a!$ET \rۨz/SW$6,䅛hMQIjXAҲWfVfv4ye y&eSR><`7 Q]d ibFgv䵋,䅛hMRͥ7- :uGs#=SjP1 I tX9P\aD<&eH ᴙ`&L)K.x9 )JCT`aRJ@2ddT5rmpLG_M4˦'9d7NG3\,mJ4\tVf;. y&eSB>VŇrv- mL[|sS8fF+;pf[~[>d@Ja4"qA*C?7W!WĻ#Fm%+6 6|^缡 +]$R^~'>. ޖaO| J߿T19{;*LWo0437r`" z?l p7F.෸((T34_:;~?ލF|"pmI2O&埘d Ys};`\ipۚ 仧~?tjgwz6T4'҉I&vƒeǞ+RXp4 \%-.//PrKOށ6vE*Uͫ$WU\Mch5qf)vDVਔGK9 BRMH#}ɂ E]:Q[]*W5袷8{[9?gtga\7|oi zGk\4Y ~$=.RєtVΝ+=G6"MD *!0( siHtp[x]2pAp79 z9zoM0Jd>lJ][X&[X!0=T [&!N)u y;}]JDZy_(7߄v4xKxB0!Bb!t4H*!$ yD*UͫhVrL!mRJW\x&}­u1H`10I5J4|oi mh)ڻUW*'ĵho>JAh$kQoKFN+s 3"n1m2ɷe"Ȍ\Krq)'&\BH 9."7`nqU1r^Ӝjh,1,lu5:p~{'Yl2Ʃ2g2ޒS񾬥S(NqFqz:šJtPڪhݯPҸr=GCaN'?9ό٣U˫$WI.rYL s2G z%AZe%SQݦ56|oaL)vRpWj`r4bZd0b!׾Di.UЁYǢ'JkAK(ocը1ʉ/ܙUpز`mpBɠDt&sk LyKDE~dYDhIrvH#!YF6sA5ɽ r fRHJґ6j],KQt1MH#ɟ#kʦv"E'.O,j sxbu &JRn:fjr?l;ړFxim~`d* -Rb+ +Ar `B!iŢ]qgO:͟ŖgNӧlWu #'wk.J̭q_'T‹%!}Aj~ T:g>@[oi뮶>?B(R7\O>6TLo\Sywc;kH/aAT":*db$T$\)"q$AQbw9ḊcRAtQ*- ;HQ J18zJ‹(͑^;L#GeA!E=+ǘBiYX~­׈j*Mz $~ìrIW>!QEJ `mG H{T{T׿C(ջHyBRr׉>SIHW_\tKYJg޾ |:%-O׹}Ujk^ԁfD@KXZKp$oKඪ:P*sPC@8Tl_aZn D?S!]#@BKpU6V Й P$]XB U+ Hl$D p_s Ӥ^,@?B靌hZ:'{fÌ,vcIĠ~b'`7g]5!tL2a"RdKJѢtbkP0Tn&sBW#P IC_(S˩#XœFkzqEo)H3G-Lpٕ]2`(f审9xLF c )p*k,\MNLJ5.]@ %hm]>ietK&&_\`"%k-݅`nz' *VnD0o_,fʤO/ 9pwvB+o濗A؄TZQ=-j5ȗ$UxZM=EJ3OρLh`'ƅN w2i.PPjR(D,LBHFڿK;=5K>O?n.1b3h/LYLMftJ' ڜhN!՚v AwPwU7gUY g8LF]uq*M 5Ŗ1VBr*gGVAThuGHrԎ&f/0p/0vHH6-f-g ϑɜ\PT:WAbE[ h 1zL@E .`˹pg2 ҄BI >(Wއ 8$=f z=geTȼ)Ģp\–enW. MgoVHy AR!l$$:Q:xᭈ36L/)x-sĩ L4ascmD=ZqgdSXW2ToCLͷ2fT罧D"hDž0덖ϣ]ŋo|䍹=bO;*6P`5wlr77nO\He9F G]Ho!'gB2o}}hYxoƼbdo -*m/ed>, l39ym4Z|i~݇h->_l0)1lw>87qbLeKMs7kNṚ='=b8m[9郳w3O9ڹn}88[kϴW0V~*#š*)p}y5 }ې ~>!5NҺ1BGuA41.SBq_%N9Pκ3Te8z=y"* (#NJ#ֆSpKJޠTjysTzc T]sՀj*N6fNQFCm]:dt׊ 0M f7^2F!Ek#Q*Q$1DbTq_,V^مiYǥH+2:,(?8fcR܎@v'*d@7Z#'aep.n_[|"! SUu r F h^Uןo;@9yG d&t^vGv1ϊ#QPxC5Th\V"I܃QFL2Cp5jNw}g%cZJ{vEQgW۵_' CrbHId\Q0mn'ЂkVP{vHH).Dgvߘ(M /=~cFuAՆK6bH~#b۷]0Y>U1_]Uޝy!Gw#8כjEqSc(2zQ7K~hILG7fqF>H_!}SCh AC3-`UH## 2 k+E)2A :J+'KFؗ*ʣ鼬q]}_/_ a>M:bM/$ZFQ^)pIܓV\"G(Z2=7:+%1Xw8ThZًj zL6$bQ3Job,7E='TS)-E܋ZI\K?)rHBEO׊^6P{,? |18`}X$n 3_,/ (3k8 ݽROD! ?bkQ- Υr)pqcFAhϣ.9MT*">h "xbt(/Y(#I8ϊX~o.-kjkYQK F,4k/~/({RG=@ΦS>}Qh^YbUdJU)MU""9#kcWSh6\J;$} 諥z ra<2 )ș77[C Ha) ֲ'T_/[YVi_(#æĪÇqrOrň|^ey)澢&1M#B\;@> +,$(%!Bcb"Fy=bq_}X^":<|_ĎXφz΋>0 'o-hz4zOh9p>M(YEoGs qd()(EhlM8TtO>Y%7[ &/DIȐ8'J&J0~N(Qfiu6;Ȓj8 t/ y!/Ѱ6i:q>6v(݈)M14DpoHOSuf58UgH:MփG|iR\ӏ✦IDuZdvIsDh&>zKH?zTHFBE-H@*Kes#':؜fYwG۹= kDK`}9ctkUhmorɒs049bjomh%L$Ώ]$!quv7^bݹZdeLE" /3yG>\!74 [0a^X5.wiUcu=U%(ldzB d"`m'f,sۧruDTϙ:H '* 9rv\+k[@|zBPڸ&̦lLrvgs5\!;V{mXy =Bp:=:>,Qvo"Z7{z=u)a\&Ni҄8|TDKBqi}ŕs`)FH0t;I) mL׋f(m&UͤV@nRJ$ȲfR5qEk8I.9Ԋ';){ݣԓIk&IfnV(x)XhGR"7-PC !gsCʒg37uI{/MߪYt,y P`1Ą~-Z!8~].jXA璠Tݻk"h +@4eqfbv۽)Q%IhX[IS&'IԷeKGQk^z|-}C2ɚ j=n#J(ԸNd ӻ h A!z Q\wE$0`D/8Sm7frjj^iO#ςۻOK>E+H#-:OY'޲R*lWY8|͌&QKbxƒ%,J8 F_ u> "x 26 -Ÿ<t SGi'4:Sв'GrhfКƩZMx.Y`1J%iItcC^V9A1I z2Eu{r뚻Ǜ=`DO)޵Zͷkh|ߊ{gd`U&8>i%vV?ڳ4ZL :sv >i&=ć섽պvYH]qW_|:_8סa:yѹE|v'׵6{fY&GYmwɷθ_G\48toϧGfabdZgTn give3gr7 ~eyhړfL،Y=W42Þ4ܜ3$8axh@):edy5d[T>e j:c,;͗֎*k]2G"`Mbt$ߡFvͶF>a8L / G=aXS[֩m6~t-x[i=eʠ/Ww]io[7+>} ` z㴷@[4~D#y8 KIIXGYt*Ec˒ w!y~T#m> }[g~ &kl?jo*1/?L7Kic|tMp7} GD*ow 8|BcMFߝ]HN 4ֹs~i웎P `s O7pWp0 GqVmrF*('Vz4ͽu2Yl]=ؖw6 O%l Q/j^)T@k Mc$w7 F(ǻQI6Ѝ(&|e0p ( ZYVKe)[ZgA2K /qWe[xLQY7XN>b0 e)e)G >9M@q/lPޥjSV  ˺W鼟A(a\E)_ 3u8Ckï;! Mw`S~069/<u:id:YĝfL_?*p ^paTnu\-.7۷oKw}9@brTF;dž`9^j19oN@+g^8u^.[ĮU2>R4x fZ~8Vď8`  h  `0>P(@,ޞpmj{K<ǩ}>] ޳`bn>r'Ր='.&Yx9?+]-0q}ſϣ#rjJD!_ \xs>P%R4KU$FQ'0j%@wwJ!ZO"8Ō]p#" E0* S ;f;J kP tw\+\L1W^/ܴ}+ @,XRD{ bcR TZ0^ނjz!.t3.ۣ4.t|c:Sʕ8ڙD;aL"&Y2+>|Ԃ;l|Q+vQ'3уGP!5',6dwR@ ImZ*ՐTu5'\{ QקHbwj;fb Ipfl}ܼ?p~aw;[]t[]igd ^ ;Oo?^l _H}fqBO◻,%aX 2twسN_߿(&9c]dJq{izEEV&WGݤ&R1:ǻuW*&[nT[r*SZ㣯nH"T N3nmLkͻJVnuhȉhN)vdGR1M|^U;Le{yfU4U4I<{ލzyT N3nEV9,&ݡwCCN\EtX{ {7bPIuw_Q{!z4U4I:M~i"h3L+wK߲X52ACN\EtJȾm‹}NK`m8Qk1zAfk9Q:~:-3LU[{+)v.:>-Y<ʗD瞭ܳU%Dc-0E=uZ#L JYɽ Z$oLi)o i *һh<('טsFK ջ3Ū5!טOLa]1S"Z5S1NIݻ3B1s@WLD:טsVKwgk*])k̹ƼN}1S&U>$טzwG k̹\%pNzw ;~r9טk0] %u1s1Ky1s Y.k #k̹\%(,ijZ2k̹\%hE{wifX0k̹\%0aw5fFƜkuZ̈>\c(w5fFe)Ԙ-ԗ3cݛ| 5fP)JzWcxX1s]Vs9טΔ]qX1.KuJKJԼKkm^~\Z|f"S3?x`8;xVl``nv_^^Và5.Ǡ_z;]# 3r3z~q|0p˷l=b9ͦ[d6}U:J|vntͮV񪺍Yfҫ-~|P_ :{_@dO<0wcmG6bqT`F(c% trg|W]v͏qUu&CK 8eQ2}gA\;Ӵ%GXz=Xl HrܷeIp$V!j̽4nut~)% ,`kX@|5;\VёKF 7qC]t3RA*T{۠$ 8x*> 1f&6ƸN&݄Nә ,~pm#I[tx+-Q2 2YȲ;?Լm=/zk{룲=?{O֑_!e6HH]a`k$k#/}ZP2,߷//Y "QWWUe@8W${癣19%I}9"KkϧbQig}ZW}YMŧMEc>pPBE :SJq-ĈbZںOV"z"Jb( ,\j"W`s$DVIhb3NCl@ЍGJno9r\LF6GM*Jec8gqϮ٢Dz{3]Fu?go?]0QiPZ},Zv'kx2wng'-V:HA=73{qg`qC2&Aٹ:(3D/_-W<(F k#?B[ї =)gp4˶Z=A5MYO_9M5gG̮7ݎbv2'R.'Stoh0&YC:?_Q\G?O(q*6voR݄/g-,d /%9Bn3Q;,&X P1S>vɮe:vY:gs,VU'π%o&EΜ5mxs󽽞be+ jD5k Q<ݤ5\u_ "!9c JGx(4Ql"`D$W["p?{^wVٮ6S@?f=d6kiԻ <7s 73d?Usv>Pv5UeFǢ%6+ A}Z8ȿp?Sl I)E6!VQ  CZdu֬+,P[Kd?yviEͶXkK Qr91_ Pط@: _0+Ƅe[H5Z ΢q8 Moaٺ҈t6U&lobs\x&x5WKRe'7*`\ ?'%!$ o\a|1[qָ2xqӾ5x|};AD?K(+h@ "` qL7AJ.`{Ax@`]62S~?fƫ!"%2w kunwZrܾglc1bLZޛ[.AX8cGdp;嵴oE4* ,QW\&ٞ:(Ƽ#޼hMj۶aEO?vw5М]tY{'SvA1-=t= >Z<~AKB@D QYdVBJ2 R>9hb*?~ь.aS%,υmW Ip0MN1 .h#ṱ [`<5ֲH \p reyCԄ(Iy'P`$ 7UsD|(J ڶň591Nٺn'Om5j.-*FFᝤBFl zvِ?}iHC2I0b- i56[Pxi`|=?1 e 8n.8)?^|4(N*mΕ(9h.<kK5FET[0ya{DQdy jh%B]Z%xb$fR9@ f=^oapi^o0Gjj L Jg#Ԑ;9= 16Ꮻ]`k`Jn:Z?=SU8.ϘWH==M4KoMʿZHys$O ̫\A" ݸ$ o4+=˨m(qK2ZgiMͅ Ҷ>h vnz PE:m5^h X:@DN10\Df#)g#(dNNد^7#'ٚ˥ȝC5'(`*d$3a$I 4u5ǖR薪RˡP쥵>l^זl~Ʋ Q~[sGIrwF?BR^;Svf_ꔙ(:E3{#FLRZMF\uOF^wcKpTmeU՗{W/8#{4ルvl5M }vՀ)qlM;CuLqڨf3x9PÔ~?UZ .w3hg b'ߋҐ9_7IyҌVQkg Ԃ`"sG Ԭ1 e()YnQq1ՀJUV&QhxY? NNNAߝ^}IPEOg^.j.mԸU>[OgASJ F<4ۋZ4AɽNz{mk(2@rڧMxpWj:%/7#(k}|QW9x V6Fd ,ixyv0h|G|=?IͿl `+ +R -w~Y+qH4GaKjٗCT+>QM5{3ajGA-Hm }3RbP!{XZ Xk,m)̎-0#|rzD2/y-M`}s,LǮ=?j@%*wv5Lڃ 0;5`T=`H(σ]&N&@@|ZJϵy]Qr2%[e PcԾɼis>qŘq, =B@,zԅ].9hI*2NZ/%*R `!AeSasЂc(%Hh jLb&5āL̢d)8 x:GQ3B a)QXInN%6a9IYOKHI"!MJDb  =,(ohCJg3SkZ O3֗ca{ ߀JK.2޹nJ9a%Rb޺b%jPZ' ?C׹gּ"Eg>D|모49<nvFT[*sX@|9;s e劎gp] qw-3'T&dƏwU ު{r ,yRxsu@1'Os?׾a4E 88 Hi4Cu3#/֪gʚ8_Ae`#ki&;V+u\u}ٸjt7@m ʬLMѴRRi%u<\41wGnH6]ILӤBV 8jT(ri9a Ѡ P'4V|V20v5KSaC{4k"_aɵ9\RMcb YxfZ"FيDq a48SSd]c*EQsufiqb/~%])OX-U(+ф7AV-1|,"AM.>wB B  E,lkA£bYș.n=6'O-ۇY1K;l^J%9AB(C!MnϬ-Fi͖T9 d"2Cd@9 Yk|dV:yl}o͆10|ȄU~ 59'i~LyӦ1X{d (.^%B3&D{Dp;A}cV4"Y?'! bPAzg GZww_D#@z_'}qsLcv8 WfGE)Ŋ=sl-Wfוߔ@Dև9xP|'.>OCɩԧJiR% ]{?X8E#L wipW!iŕ}Z~3G_@pߗs5-?d<ݤhG\GbcӇuGCӟYA3A(H69av%xWgWo=lla+vLsT=i>˿lӸ]S&tJԻ0OUD&x&sը Ҟ9^2LEGg9Bg{u0鉮*{C P~8:#]pWX}^ Z-UDsPZRI#Ea~'>us> c } C:~_daA`7AƊ)[>!̺볳eNNLNo*- i-<;E18< SS,'>k6{c_:'sRyGcƳjR P{'4سQ {Apji#Nr`$S:THn=DB4{[RIo1I-sl̇e4,_%?E?M_ėhPE4c4}mq}rRnU;f/pӛI}ĂF"n4Hkm5!"fవRY:gZO%EMj*XNt9i00Rhr64/̯CeO7o<,4ldYllS)z6zY&jJU/~qU/:J1TNˊ@I^nBi@IQ):*F*A]ruDtGuMJEw^<(('/wo-߽?uYwܷ32HI#nMI -v=m8bǨ2\ADzv{! TxUQ(ăqSҌD@\!f*b ֪Aa!9Hz:"*w@ΫM;h|}'8K;[W"-o_._՟5- L2.F<:w;+JԦ300ȚewAWC@ \pĆIx5 *p+.&<.IRs,h³X[p#LIW:˾&x,:. v-gxbdh/Q>4$֍O͈_#vJ̍!t"])ul| > Jͤf,X@xjh w:h҅P3yqtl>R^YiӐnzSL./.Hc s`My͵6,z ;.(e8&!U-rUҳO6yYqo{5 x"%UvjH}`؄5R{75 )$.JZJ`[2`ˀUCI|`%Bb~7ӋWbr~&& c7bdf1{ϙMeyRM,ӕq1k8:H!~3}=t%]KN0Ŝ?,eڼfY6ߍ?m]z4x[^/>߳#>7E5ZmXQM}J*Q5B}Ct:']Z0u`&we1 LL:<_4cӤ1κuvBok{}{'Mڲah'w;h&%{Fr1ύ^IOe2 [/=`/F+TDWKp4ou#7dqzg?{W (5 Ɯ#5m[^߬?m(MT6ޝ.8)L5$%`O )&*"V$L5(S88s{tJ.'#m[ւVNNpwh} BQ9]UA|HUJ0} (G%[zp ׀ PJKׇ#h5P.Vow)A0aNj @y #v=˻7%VR֛ȫc\6"oIiƓu/__B"}5h7b9EZ[}{Stu ;]PF"52͚HC*,MG1 x+S E53:b @j04rb0n2/1vRʑ`2J׀j7?OghO/Jʎx,=pziS)J'$ҙ?Ԭnѵ+T=4/^/t3^~t~cip?Ð>d6??lc鑲+lЧti_?? F~[y=xa_HO|LgL \niIB"DX%sI$gnMy#:Mc'nkڐ?6 s8F?֔9t:Q`ɬ[Z6$.GS*%F)4T2`(O| D2I4T)E$:m3cΕYj<>'Gf^yJ$N+KYS"Df6p7.m;`zJ3R ېT9fYHX; !is r O_PvJQjM6w+UUSmp˃&<O/g@C to~*W/'6/v/nx1!y')BvCqOD\)88&?y3e>V= d$1P0 0JBjOM$Ωc9*!+o9i,PBLZ~&@ҋ!{`΁W͆y= p?=E3q^>{N{ EE|zNh}p[A4dK7UrοWJ䈼$>h*iE?k WI2\%e*C}b^Y#2 Fȹ j*aKT[J8otF탚N~̧\FŻ/y?J(Ҭy*m Z"!K[QDs,ȖÚ,oQoULޤbV2 (TZT-)Ӫ 9V3jI+%]j~*I8X$lH QY{s}걀N#?#9~.iIζ{[?)8 G ˻>oׄLos[A8\hhބ}o1[g2L;&@iLgI/$ZM^|X$-oY=*Ǡ 4MDZ>ICk R..靘uZ9H_T9CToTmuľny񭱹d}ΫVvMFY3[Θ+iՁkmF/wEk~KfKHrOl3+d֛߯ؒ-v$[ [-vWbUXuO%`P/baZ!;M!h ocIoy)Cq[+aM̋&=kݻ꯾Lz_զ SJmh @1B'ȵy+ן읍$> ..d۫Y/?;V)*y^Wnb/gxH)2ȥ%*zgp>#޹V:4(8~0t h rhcZCa).) gp^ȽU5H߱Nq!6CF:'4}Tvd@>|a[njV%De'yd7,8k/eoRJjRq3 =M=,-'0̕ᄈRh!i4GA .5;x&96s"g#aOTsJbR3MJq5IWo%cĒmH;уeJaL0;Z!F P{͈RJQVmQѼQQsQ-($0kL%r%N K἖!f a,J~ \ws t]vDфM*-c N1p.ڒAplǔ[i`Ċ`F`~-UaKJij}5ml0% r?bv{931;jqj{Km=BbyjA+kr*]f?Hzx[@BD4bw{K8\ғ>fOdфQ~Շzk 2JZH1,jvlO;4`: jjt|w㟶}x@!k\yp(d%=1b۶02TR*F:܃l1c2?q՘aVi*--WV&qu֤ Y2h- rXHdA lOϣQ\eϻ1oN: ̏łM71G~H{8y7>02"L,OUd^/y9Hw79ŰP1'-nEb=WqZfXb,y²8N- i%Я땜)$폝 6/}3L ib=繨؎YI^<}9w0WRܭ:f]5eu-PzS2!]̒NYgMrĺ86^kq4^~Y46.~SƔ>@]: 5Aĩ@8IA&ڍޏ)ƪ[V)wSt/%W ᓊr^4>E=v@LH-3h!Z#MI/p7`BE 8.eM$[7ods` z2--#Qr&-$^SLA%ԕS+тKX !M bC`=?3x+dB'[I{9-/?}Z| c㏋x5)jѯw`+j@g BGd.pBXi+|{pWDAF(ѱ=`d&ECé!R*h qKW9???d\{ǏQ0ېZB$x`l WVn 1NԒ~6$xu+ؤ *Ɵz\(Jx wJxm8m4 8MD$ 7\߱8؇f"uTPeɞ/<:Gcݜ]xso`kghŵUL$x9")iA3 *Jb(8x9NCqCH#fNs$4KU dX) ?mqy@7=†oZ)j)DEL]foNm@ԛx -,`˸Vؑݽ9SF؛S\č!L ?VV`o-P6XT/}.!ouo~j;)ѯ_zrak_~s+zz<|)'_ӧ0CZ}h%uwf4O'*v)'߫ 9v@}YW[׮> 6B5z`D C/uY#@8؟y4uu\e^Dב!EZ#MLJqJ$nGojڲdV߿~_W.?0mbo c:41̏0X/UaALbΊl!6ICָs{k #S +(UJ)33x>H7tR.k{IrP&O/뛚QcQ-S|8ՂvD(y-0MQMHUXv>巯<^W_ V &[]5/IZRZ>R ދŸ JD}kI=(Miy3k"†SÈMmtԤ<9%a7Ge [ޱ8\ ^`mو XlM,vo>K*H)+) /d82v>4GA^۠n E% ۢT΃hV5cZh6h.dd|W=D.2_Ѻk}bis؏㖁\P+q@gÞ W%p㭉zWus?|_$&lyCӶu[ufw[V/ Pճ}9;H}=xYLҕ'o5ƜS:_z1^0V}ʾCk3$;4S"-ŬYZUtb4a>WfgRzG.y[r_-œr+^/d\;.zxݓ߾TuK8D3aD ;*r,{%?IAuIti; *.W>]_~,tbk[mo~[Ǣ4|0Ů|DU( !$-Rcuq9G/뙤Ǟn E׃_ X\dWoe|`^e`AV NvsL+%~`2o(ՋxmxmuX'UpaJ%{#5 GP))Wa!Ys_%չ/_>NbNy3Q-%œߖ1*$uńjMgv8C^'F*0 Ma(baG,rxfeeSDN `c"Q&Z%yRrɽ2yb|J*2^R2C "㯥ВR rE&=Q)j՜i@D-ǜgӱEgݞbˤOͳ wY/TOQnHE L_5zwgּu٘ ߟG#0nS\Tgd&ӿ͛~q|$pmbxJ9"\KTT)%PZ;H:aMyv 7YzsfK@P4g TezH Ѩ,=Yq}P~cO$Md(,*㮬6KgxQƊ{jR uWnPZSA,[P`yJb, o (wAYu[y8` "# IJ0.B[*8)t`d; ̕?{F_wؖY$pd fnGYIr&Wlv[[-ēA`K6WEX>v`jG=]hA`d;M%dy$$[X1K cUR{2]}O$j}< ꍕȍYV ̜^OIfq ޭFTuQlؽO!̢q&KbҹR@`CDeLTdN@;[ûuPKhXarۆf`XϪX |nu|\̉.78YLod/}\,oρ!?bs ȍ ?<|L_Sf1̳TrS\_/:,E^`Y6W-xV2׋[Vr$d,lQGھ|,Jm!q &))tM>sX m#&sߌ+ΜyavӕFM݃lKv@|v{J]gb:H#Q w`^LUaIm؟tMsHfh1- j2g3XpRD 4+J0*&cԑWƽ˥W-صpĈb X韬e>gY9, NM~a>Mt8 Xu^P *"VX^5<@uTk>& 微f kjNTz=D2Qɋ4ҵcy -k®| פ #k ֜ F SDg2XdQ%|:xtLD9>1+y,{x:g }n\s=IS;-bA }R9K8/]xi67iɱP˝y5|e-E1*S6Is 2bhʜ2 Enʼ3VQC`5vFs]޳|ʤjD19ÂLmS2XBTZ-A.ZmG0xc}.0@>5 PX 9FlF<}a4Sy,cQ_3.H&N6E=dLYO!dtpi(ga3VFy[KXJm 5E6F‡0V]%o.H@%|b4 K/󇟭+-X#a!mA4&[ri +f&6ɛ`03Ŷɟ/ ngNLeY vmg>]_Յ QΛsdRROmtaۢ*K:v*&YH-}Q#t("Pxǭq,%_lݓDgG(gWi+oM+( f i(d]m-vS^@6#)0lOgc \F5Kd 9'4)TI&4rN6.cз ͓j%?Yp(r}#ܧtPz(N|Ņ.";H,DV,sR{ʠ+eY Y-DGrpv+;P"Mz+0Z>.>|rRv#+i0Gxz5JhxIv(%PѤjV.:&4{mrCW?Rs"-~|wv5ky*~Gɳ7,a5K .iI g6Q-2Qe/G*ah6*ߜwv/ }_yPoG;Mn,DG CZPBChZN^@ e@q3] |] ?Q޳YC!xʥd)OtL!"Ζe}hNUkFpdW7eoO7';pK!6흜t\>8za9u3[k'N^?;'DFB+ѻBAP!Uj!oCG7Ptf`"h`m2vٙ@ yzۘ8޻G跳n, N1O< H/@9ZmYD.OXm1_$Qc9ì1;T Bw韧G٨zz[u5~o*wjoV~w5lvq6e߬~q1_!Mf]s3{kBNk =AAwq^}DVD\T*xnfB^EKRssCV.1HfRz]+˸ry~ EL)|'f4[W NwN}Omwgpw;|"EUC Gv4h!Z=v E.S:wl S!YVweIX>80.Ktp=~~Ճ|w24j/?xT^}F~ջ׉ y \vߪ#uE2{֙iXݝ 뛖RӪx0dKf4D9VCʡE%\U '#b [juNŖhNLeȔI’Z!SۨN,8:9 V&=Z@+L~>kDp%v=s:Љ\"nƋZ >0ibׅZuLn_$uߵ'gc%Ң逦6H#&"_tkOWGJq78_cVo启lm }(SMt\GICRޙZG蘴G.Uz:e/e i!V$cEE3˭E? }8o2w ,W(YcQi)qQk;Έix-;H6SkIQ/o3TQkG,=>A LEZV_J5;W%=$ԒIšks/ks-ۧq\8,xf._[lC!C{#Emj%_J8H#)0G%+zU5b2AqkOg*VGOq&޿1_wuqhD3jA 4h"d(Pt)O{LjxNʚr䆖޹e,} t2Q͸#hN~yoVTrbobX(s2*l͵unOh3trz$GЍylс>aRI͍h†PH&XB @B{| q%s.7t7F[sD*~:xD73~2YmJTEMi=K{}Em쏻8`K=ƚ{?|,JIw u,xp:))ֿudbY8sk:u|Ǎ&l3IJIҰfqIGծ{/NKt3uvWm,/.M&Kå{-kD`Ke0ќyx6*47~eh]bkk5blo^(PZD£3B-A`?dg=Dag&H>uq}ꄨoэ@Hc+(ED41w42ov dMdTgRZ!&1XEL% L`52xzW2;ì^O MD($B[$q%.NXRsT.iEa0:!  a`7/OHVT6D"<-4%( hU, 0,W1$M͉ۚ҄a [~~ȷnU6\m-E4Z%+.`h*:Ѹ3R% +]c3hfJTZ#DPk%z[yTC@!L, nJKzd<=NhS$n@uK.eN`g?ƃ»mPNI%o0af$*\o!eR @e;e~n]oGWpC=\kq咅KERIK WCH2y=ے|UUUvg!HF 0&Oe`qf;x)BKpxY Ϝ󽋝+K$϶B^h︔ȸŰiVFZZ`+ \ `PM}"F@b9*+Ġ*@XR0\g;*͵a([H彈Z)wm4͡G4Pv@c(uS4WtSq. s>ݞo_59."q8w]P*ELti#{8u5g2>sk܃en4bO`SG#=v81$.И94& 2DSH o%b /1V&Rb!q $7:s X h1<.І1VT3f&Y>2qt}qڇ!2!`F'_fvx~~vߵ0[t6"D112ջ-Լx}Z4Ig !}%oZ6-C83x~|:r4@\w*"HPQp1*ɜ? }(H9Hrڅ"x~9GDž2ɲWd2qO DkPwE0tыuD(]q87JU Qd $Lhݜe; o۹]}ë3Somu& 6bgH),e;TpTD ̢ cSmPc;0e$#:S?JxU2yq94u& VTEh U2td4b6"yn `79tIu՟K&>,̙xYost)}z =2륖A}GqCCD%.`<_fxC[DkLa65ŬX0v8֧Xc0-ȁQkpiP(t A14ypRᬑa#M`{9YWkR'Z!$J1ɨqL'uGԱ i*VMiJUbMQQ ZI ?\!TMO괡 c8o=80݈ֆ )ۻMtUvҋRI4-"sF +2[yؔ,aYòFJly}EP2^Ȑ#V` c 0x6+;=7wT,H/zt4י[ja0eBB0_iPuVuj~9SE@a!klZRw[uxƠ(CDr-S dM%r  %k$ˉ/fd`njG$s0}y*T U)Mn 2&CdȄi4W5|$H'=52C ^ ҂e(eGC0Ɗ6Luׯ1'q>bUU;J]З nvh*jMe]&sSBݺSoе 3Ei410F` 0Z6])d1("uiނLmA5QrcNܗmO}"76^.)ܚB4)oBya~댿|& Wi|8rՙo2xIxHa"I(H6"͞]So^1xVsUc]"y)%,`4WBrґF,1l|Ϸ1oV3.syUC7,$Y&9ST?ylxx ne΅۳&rÉJ};ɝVJ=G X?kUQz8FOBM=A<(s\by}chp\{5R [ 1uo$`|c%mY*RZg ;41Py@u*9UDVM}+Q X4 3QC;d?Lx X2I^6Pa+Co ʐ*Ua05iK/9~F.ky3P!lX ɵͻ u^!Y\UUHV9fz ɵتBr]mU.%ggf8I3ݨR ~r\W8,qH7be*^=mOTƶd(nX"H} U@ &SkbFt#5Vur{[zYzJ"hl16]{u u5N|usY@JpwEr̒vȬp)Heo@Se xU7pzSE+V9 ʙW(; P}d#3\h,afa>qH VY u~NDL?s\vC(ӄ EtٮaH%Op+Dz3$i"4~t 4~ЦaF0+, e/'"jkKa/X#̺|+UlcRH*q;[,)V%OO ):!7ݜųɳH&7B{<uaThViGb6*n(3Ji4=jF%Ѯs$>E0YSr?r} (0KK{*Lq`Pj>Od?@OX'/ wҟ`J 7#aEx H-j-8Oz-",hl ~-Ѽ[s?-Ӄ8Znh#]:[Ǽw>Za9,$`&cPzFI%d}#y2?#zF3Jh\([.߀ϸ `N1%Rq8v+?m4sɾ#XyXIf0QN|#. 8wq$ ;XS< 1LU(P$.)"T"W» MYiO$0EBdȂ%sƈ6L ]꤮]ouV$T%*"鰘RZ/E/h_?.E^G?Jwgz#͇HaVA- \0B1-AS,r$  7ơ(O3BY@el"؀,Ѕ g=Xkz^a? hv`ë)|bKw'_ỿmv>9/㟎=y}t| /oo 㟾_wb<[_N~8~>|VoolYi8z8/>G8u}|n4)|z9 ˟P]oGW ( 31nkK_ރå(ő\INʒ%EZ_RJ!Emo_7[6e 8ma;ymL)o\&.fm/]^=s38O썇qd/P/]ыet u/>*bܕ|ppfKp|S !~z18OǗW.8Yc3vw̺)NWWm/;/Qz?z|a"&t? ŕޞG?D^]4 &dY|h'ttlzxg}?_eK/Md~/.)v?D]7ϝ1.K^!w CK3\Z7kŒDhJH@#_v$@($^_}\õ`~Qjr1N\N*ч^]b4soko1ۗ/%_dkl{}n:K|ϯ^~]e]5XJhJK?$+RX4Vȉ&fZL(T+B9(<]7R`"[;u57BWX_m*泻2P}b |EtBak5xm8 ZO$ Wda3h:JJ r&K_vR7+TY?=gƔIFjzG *ȣ𹮷͛]V)EH # 3Ҁ 0@*;Y]맓~wu~Wwuڜǵ^g.JC&&|(50nO/m͞3_ ; ZhMb^WLwdReLF׊5_^k^T 1T^U U~Ѽ.'V~j ZRV 2֐i*c-Zhi=^So#tV闆JT턳Vd"(ڭ;>HN H;,4a$'Q;z@=f| Yi8gʐ&PMA9}DK-vHI"]d9E]m|EbY"]dqsƠXEcԤB+ }x"@#9vѬ'e9#(d>YǢ Q g TE*S*IƄ‚ϮÂ`;,S^7МBX c HJH(8(RMS0DX9c`\yXJA8HJiA(C v`9>v``7h JF.<S 6$?TKu6ӯoU0<*ĄSGˌ<  `WVX$ye-e(UNaKKJvg4,k-p7߽yO?7xs.xƴ% /_zVPJ[;}utuz?ڳ/{G% pZ-Zv;( WE }HJIJ{"۩Q-zNϿ.y!{F&n޳)R q ]PHTyI3UvS<ϿFꌤU"H>=#i?Iս7C5z侂^쓡 ݒ,|&_|~*S 9;;t=o﮵,R`dCq{o._fL'f|*ס$Sm4mέ3.Mb'/yO$gkƕċcc,&r`4w:1Da9?DPg Q! $QD87 !& /h^ #wULJ(4&" ι :'s/z1)^FLQM5_*8d1F)8EC6&YFI՘%'|!Yż{ Anf)A)3ƣޱFU@]u"y]]屘XM"6"DA5q+*k"KHD.H *ꏂprb(e_(14\8h3 t]u"$. J8ňwK֑TШ(x4bI# `qQPKZAZPyWFH2ա2  }yRdߖY[X$aN[*TgA=+W$Oa~Wo2xphaڿ,_=79T"_4/I\ߐ] L $*5TV% P^ꒂQlIN9쨢 hpyFBX c R,=ʽ#iR 鏓9M[eʚ;l7a+YNmK~V <@d͕jL}M03xU~ ) 's3Dm-̴O& L-_,_{ T3TFѺMܑs,ZRq`dC0͉cHZZFserj1HmI>-:TnKImIb=뎭lJ-& Y'qV[tdza$`ȧYL.y%O花uD$5sflC?f̚K'UPT]7 *CEZkqƁ%3 # 8,|n8/C}JT2DmBf[>dU SqY.G-7]2a5?x2N7[2ڻ|xpn2xK0Zh!%㬳@'֛ 6g"ZBܼ&RLKĆDKƩ ;L. }KqZFO!|5Ͷ=땦w'8Ɂv!a 'vV1Ԍi𮂓+w\)9_W|>l$P>iRbQ<>`$ ;yr|rA+7_R!XuJA㰗ʖ?pN]nUxl8ȭbxAJ+(TJ,Q䃱e)ts0tUYTcK+pKzjlǠ ˮZfs>SzmZwwgp :Uij@7aƁ=yF\wFuQ~N7LNBgpSI4{|զEoiݙuIqx"yZe{vyZ&O oyDž X}@UwAR8%]Ӻ8_i՛&yi!yJk5-vmcprqXma'8(c[,zq ;CI"SHJJڸb;3ٝtZ Ԫ`D=eM| b)ڧ@~;IjU9g * rX!mX,Yi6dj P_MVbV2e9’OZ3[@J+Iė\w/uhefl.TUC뇻˗F7٥ۤU3|)q0u, at^\ES\@?S\@?]\@5AFetLfUB)4U>]1G61` dK0O#5'm{:;Z[(n5}Qu2vɨ 9ۼCelKKұf4D1e?4?s0ӣ$F.$c$j@5y a+QRO /*>F q{fv}4h<#T8ab^6*-#K@qajir<^o"3Lf7 Jh!g}7Ę 5*}qy\d'> 3Vc'Mq-|(,ɳ RFY 54[@'ODY3Frͷ/65].(lPr~ף)G$|,wOYP(gNP;^G%8$OHH1u(!X{ 6&%pf I"tzn@1>Bק!pJ慯 e4ց 18yv X9G%-ZO3սQu/Uv] x|&GDCeaCF@'-ozn4?".e3mxFZ(O}h +kXKÝkT!ٽ;v6g-m+J2K%%6TIEY|IaeY<Uf]<["_)jZ欵pCEGר 6`HbA]6.ݺ*4Qhl7z.cZ&([Ő,[bR2r%ZLIfԌUbBmVT Wb Xaa%.`+;644yd-q]_,r7GuND}Enec.X9h&II b9vT8"uz S%f#cwrs҄m aC}^Cf6bPG>0zMpbۀGT"AµZ\Nt&ɚ̙e]7D^D`uo5Ex 7 Vh#l㕈9$&)cɫF9'2>A` 0S'R-Bڼ P =+~T4e(bVa|E 1x*9{K5T2d/5aׅι>&C" '2Ul$!wq1k-*[PN,ךnX% gS^ј L52M4eEܮOBmEj~쮺BCgˑ/WUU _L4b_!un\I<+ZtF (ͽ^A pPv}r7FΪL$G)F@VlaP֖rQb1V U,͋, a+X@eJ5Q2&DA J w҂XneŒ2)^LScH\GI9UọH"a8 uI t<1&2*w_ݵbk+6#<\L2;e:D--MXgАP[j-1f=4ZJ9ʣS ^3O6+O(p QM Oe:Ĵ\I٪Zn" UIS߷nar{0Pz*@. 8T9kV Sٚ2grp+)߱QÕbD4LFy+Xg=p{J?qIqؚhS7-_D(Ekf^)f>rs:uE/ ?o>[޳i`0μ;qj_c xV$c\'tM, 3,st5hqK kJ ]u _WׯAXҸiSe9${g+1ǦȮ !æ>wUB7ϛpi;QscJ`)nuY*)G'f9VQf^ pA. '$i\yIX\:Ecֲl-RܔI.WbGS+mXS}8n6ڔnj70<0+B1c :{K_SmعxL[֨%,剫$}7S_s!;eꍑ!-W{Zc#$CS}>EĺhOb6Y@K(J Tl,xJJ%yCSZjWRj)3l܂i$)-[&BɦS58 -6to]o5_A51Z&lGm {5$c+lG+QF-3;erK (xY2,gAI*/^TtgqF?w\)`y;)܃Y g_.۳Yo>ڿeY #шGZY6*YaHyB)}z/{| 9 ?_AR9w|7܅7?^w^tN=~7qIls2Hq =syyzyq՛3{z9~p7/9}o>=\o?/W/ON_v}8qcŀxt{Lݍ[770⭌߶n mm /ZÜ일ݿuKPX}7_b/$~5XK!׼5ōjZ.o?9/Fue\x]I/~~gWn lkulyõЍ>2ɋf .Flm%q{<pNwşߟ ]2l/^`TOf^x|͔T_*OѸAb\O4GN_|~ x|rWzO0+@Wc`2?@kr'bޝf8T&x=~Լi8~|+\QMru7E{^xh !^.05QnJo qq"twA;0nܼt.|P^-^ݼ`{>io~ W?bΥnekR:_~yf? l34T޽37*ޢes#_L=Ş~pq=ioGЗ,:^]! v bBb]&)[`jRRlJnR؉Dz>NVʺg_Ȳ+^#FW_v~WhV 1K;<.K OܬP]+j@V«䕓:u{;!j>?DzSK V +4$)e7uT!_ONqoLn4 I\ чɛg28Vc 8=)xKGh_ۇZQw& _B>S:7rsQ~BEцSUSaZӪy~ֵ->Cv<\QY7Vhf*C9u0jYQZK&F;y{@CLx\Fo{bZ\=tN~:zwr| G,WI=2FHBho\{b6-V(4ߚfJN DRkLXiequq^?dcKW0~F͠Q"To`Y)P|.j,G7TVJ#Q8ˣH\h(f\`u2RXRɓm=/WN/WN_ӕ˕˕˕˕˕pWN&JҀ1{?Z̃a*e*0Dܣ'Q**} \ cpcO|nkPhē9aw3pⲖ/۟3o?tv{h-+o o_߮!xB]΅󨌠$,:cvT9 ZN!<(T5Z=`:GrBxi|v# 3& 5BN6!܋ps.Fm/YlJٛ4ܸrr1|_]=8y)k/ZĽZU69_޽*^Mѹ[3*DYqbjӸM8Xs8X!f_-x/,q.sc.MtYU+\1rt:O]tJ{e}]W_/ߙ㑯ԸIp ravUqVmY7U[!Z%6 Gұ虳(ᢍIW2I,7 %w2,f/}_3gm.P6lxRiJk8tpIWWe2FvA')CD/D@K'AiFy$ Ry1cԁV܈i'kGhl4{@Ŗv@|Y.C2& ׌.r  $ hRKN hP=L9߶OvgC)Wz* 2&4Rb2 5Ā.'n0JnFSC@8ywM'hbgwQl(EƐ<@F YXIcAW,XQYEYa<*8(r HJ\QH`3]28nEH%Q"r(*XѦ#'d54]d$QT{.Sa\$]BFhjh I tMAat#Ub 3oޠuQ-pMto`޴&L@IΔ.?)1\ZxR'|YRȓ05 L OdܾRwI귍{Jә%Wm]t>UO8HMkgJPGmgJ i+$%дKJDtGD$̕a W :BلdgAoۢĠzw ]w-tdFR㛍;Jڨ+H,n|zD,qDVm (Q^KǦSgHzp'$ʽw|2BoFػ4>X,ΪﲈkŽ!ONE/|=nQ@&¨g14'/ +%.ni D9e,w}e}f s( ]?lnoD Ԙ<6+hR*:It,*2 蔪N(v2KmI%ϙ!s޶E%RƢ(#Ӝ,p-Ϭ?륣]6,pMF߰<0p<ŀ.2)+PVTZT%ə*T3( PCUofĸ@YD=ַ:\JkAI.MjOp P4C9#lO / {09hT@ 1(LYgf.< AnϷR;2\yyAF?٥йAj.. 礌_9l]׳<ʴh| 9,mMÅԨT1ʓ2u&疨ZtE q2EImf,' oɃ`ٙTAۺ5;I2*²8oNLtNuvK\LܶE0Td hQb*{rH԰Dt= 􁒗t I4e8*o`UA7q㖩Qc؎.4؞'{/-*uf갈g`׶>>t2/y-,(ַ΅>SٯsC!~_qviKOŲ*5+X>uf/_IB+& H]B6 XGfћsTdF#i@$_]+/W)̖wV岯ά` }J]_Hw9ErFM 4ssfqrqZ9{\X\~25%/}ѯG{x$_N?Q>G# r6<ψUou*wxɆ5oѻLFoYu,Զ!eE>V)Sey V62ZW^m8:],릑2G δ=mRwæ6c|$Iϲyy|vco]nÊ#S'| r=}pl Թ"tEmul5|$uԑ}"lydz'$ލQ!F]Mնvɹx*I"5g{zb\~MҒ &աJ.G*kpcfRx0?mcY+FpQSC' Ҟm`i+l1YJRMH}nċXQ֮tх̦,YkQ[64#Z] &q9B?R{c^j ,+DgE{i8 U?^fK'JGli nI?C 1**yŊ̶7]!Y?\P91hF1ΙK㜹4ΙK㜹:g1(PNJ@%2  c-QkJD)∡0J|BD1E+*sZ'7ćwrN7cA?YfNbsbi8{Sz7f.q;o Bk\ƠVWE^ߺ-{ ~Eeim!l:ۄzW]0,by{ÅU_6F'< =DCE8%j3-e&6o>BDmműbcR!>@ߘ&zFZtSpk}e#%[7>vfoQ2S}eR+ 7 7W1x+7~ \=^w  o82]XTt(;+Y-P20o}΋㜆J%M%4oۮy,hγGٗ1zG,WZW.]V=6#pE񺭓c/R;h v,ݺ/N]WA20[vЅ< y-][ lhm v pBJhB^c&?5'7-t[o͂eoZ<d -sVOm7d~D t{LnE%dkV~6K:_*\~f7}8 b\Z3D9y#=˳l4* y3vu)$Q܆\B.lT~ٮ\* 3/k |k{ߒ/ɷB"Sɝm##Qwm1J?@ #A'd6{ڜAVt#3ǰ$!o&.*+[ldy]; pi| *EUmPRcH{PǞR8D|^VRś[`fScR4kDnٺ!PvWBYgM0VX+w~꾻Ba*E} nGq8ݖĖ:D A5Ŋ 5oO)m[B1 6!5 c#kL54>ݾVSZ#<㙀A[:%=_`i2hʲ[~#{'W|wTߌ8Cvb.7;Kne9$Ȗ 4( o+#X#0*Rb2 ^?`sS"PJLjK'zCg?am p$ y}Dx}:=W?J>oTa2:(2 *2J 3$d+AِgG2)_7Ht%"H_˶v::"2c,)t!$wGI!9v'elS- hWa wd w0+Z3~;ϟ)!] 6]2Rio$+;yQ.àIw ymF.uQ`/ҲQoĭ`ҏ ie7pں6!OiѹөҌ>>SY 8'ZC]go##WVYYc*\S &RKdqE o%$Hċxh¿K̙vhZ#!-Aas!# V2pQТ=`Y:8D5'ȆkIS2rٶmC6l 6ƩZ %MCd5?v"BD;숱q5yrۓbf6L\КF l3r58kY@c2ot"pŖb ?20XBJ5$ΪhMlhj|LrC5!<¹ͺ; ;CKCa]pã3!3QX5ٜTR#`AYM✏$ڽv*"Nq͋$UZftq?L )P& YQ,[* w!a) +E|B"8FΗ"b3ąB5$V8uWGҖR3pK21 =f`X"/khRØ.3Lؔqء%rѠbE A]g A^`K7VWe/;}c_ fj7z:FR Qeg w[mն^[ZM=?,5(1W :TR/bZ F.ksnI A=2Qn#sV*3-Vl'e .1tNԹdX^(st2^wk!\j_r+jê%`KbR"y( +{5A:ǣ@BO!{8 X'ޖ:&3ͦͦޟ^ea m{ٵoA(90,Efg3D  Dn^(ᚮ#癅}G2d. Y,`RDXv!욜{xSD|Ү6_[ykdUbyOܞF 匸ÛL+B .6Zpg,C ꈧ{NI9@c:<&| w< ʨXc`26MCKo'@P:n7lOCiM0XS|{K%o!#UV"l ! $*3PA=)}p۟Om(RGs Ҳ^2rKW &A ){Vcz^S E n_URGzG`?y܊ *M-Kn XT"}ި+:ncp0B\W&LMG3sD|sZě'ȇtSDqAWm> <3N;"X.yd)Mo@Z(mk!-uӦ[oiDؔpkmFZF?neezޖuMHZe3cl&vQg;`8FH%r@eծ1 @R!o/q]]{C)EO`o@AaL'xx$F˓ LBdf_zߥ]pu.1^/c0Nb x8f@\1=|(Z3q`f1˾?9>dߨ䭁`IwНe<>}ǧ)%>]qcyO:O_&xI4B ?L^|Y'i6>#(̃_M&.+)Em+d~xs?< t{K CѰHI?Szݛy!oݴf3ϧ_!/N?BGñOn)??_u|1HN:px2zƀ|~џv.y~ד3hlPߧYTcaFhwfQ?x>4n;]|ͽ:Ywˇ{SlJ.B>{7SOP}.4 _ px8ґt셷F>s'ݛ,kZWsPTs{qrjE2x SJwnwaovk{]D'F*gc+_#?0*ٻW(ϧoW_[>րA*f}\02Lfzj_GB?ŏ)'8G?A;z>Vv 7g (}P SһAo'YG?(1!s(4ĉݛ7֕Ѧ_j4z5O30bW^R=Ms54o{ƥegEߛe[פ>ODjAKܦ-'Aj}zBaVjq+H"ogW0`aD0c8)(>˅131np30 O LPZnӓɗhwsf"H)zlqc?ps۝vVpsq{0Ex†"jY5PZc5Q 8X`?;٘ X eaý46`A 'p-;TUe|ݼ:n,񠣡htLIAZGdQar v/edJZ䶵[KK"Ea}׽/Љ$YF!J PQ-U3 }8U_arawJE%0 N_ ,"d͚X'?=R:֟'U%$ҡH|6.?z S?[?04XRi:T* q`Rqw%HRcB0g8 V(T4$*$,%ts=' b1#09[N*TPs%1pCMVcw61]`|u6gG^]f!j[mmd@Rڤ|H=9:QЂ~ޘFSy^)nzHIeVk$s`ɥC2@}<ơ+/>? 5)B(<%*i$e99Z"9CD+dK!$UY:a^Q$u no_W/ "+|W xe-?+EwЭ3e4444iӜ x Y-1LRj+Pep q^6&Zt-.HDnbR5$ qrF41ɦdg[DFC\ d  "z 70Pa#}C |@K-:+ySgu40=E?z~˝mSH{D@"apB_z& W'I`'#-q*Yݹ~1NF&#N5Z\bZ2*\PbzF]x}M;Apì*.U9H-/2U: !gVS回.%T/vV\痸@m\4$ZRޢzc7V KyT,߯^] ACc?zٮ/1fpBXu΋4a7?/l[q=H;ۀ}\Uo )pwGz*s.Ӌng>˝=n$vd rS4FT 5~'NԢQ87F({nݾ%FjY`D Vƿr!*kxDgyH ߉NTOVaAh#=$hRCy=q9 ǹ]miBxmPK,m 1-!94Z01~ãCTei?Yut`k&n %k >rFO ;vXߵ@?QRrMyxŒʾ,Vw&FK(^?&e'!?~|~&樖ݘcZ.GO&#rǷW_ף ;*9l6Wqueʯ ԍq"1X5!mf|;RCU8WgSo&I.΂kl1i :4U{|}w#m> ocJm0'ڥ%LCɏZ5 AmX>L+hӇ"1`5epaE-Ee|˃YHRWGϣbpn8ξ}p5A>:){hRd%ItAZ4XtKBikFҭ YOHoMR 2i9ťHћ%gF=WV>i 2{ 176jD I`cvİ (xs"E b4 &B4%)Ti-RgYq=\|œ\Y S{>$)o~qKf_->r:o͕bwg|?}KdULa<-5ǷwsIoT,o\杮_޿+Z_T0qm6yL_ dY}Kħ}1_l)^2ri.f -??tƇ*Z܀hOb7+fu΢l&X6UL'ɕ# IέJcj}0AUv$6Z$9[ŲbHyoň݂&-h݂&-hw J]{Jژ~leaxEvq2^xI$a"$%ЉFhFndyQf<#iىP|u،ɯf9;[c3S\\vkSlGlfoCÑR74?LT(etz ݼzݻKYoAlR㦟_׏}Yj4\4+i8I49sigF~*o͛ann?w/U`^^Lی^_oCwEǾxy>˗Y|KEˋKN~ p☥b8R8eP?k ߞ>M͊KX4S}ʜ?uC܎"#jraJ{ɧ^^^57/:˜OW{IWzqUPe-m@Kr퐺^b9j6R>Kb_/g?2Ь.f1>SFf+ѕRH+]J2d&Mѹ?}]Ij_j ToL #r2 kX EMFK"n+m) ԇRSr怅e579+O'/a`ı^aDE>u2yGzI_kx/xһ>qD:[8mA) @Z< ., VﯯPrV˒mҬBt䖓|;1z,k d';ς2N1簾ɁcRNjX?Y% d缾bQsǖ%Dw"@=]a(yw'E=Rθ] ="̘]BĠ=kZ e?\rUsZF4T<űs8-KZ|Fkᬿ$U*zmxzQ]7=oM?2sFgN4 фuOS&'y9\@/F'锚ۉ`5s ޵WL]2.OJݗB{&C'ez,5)x 6:ãQrvgݩus,% ;ja'lH4[8uAI"I$=)PuL>E3h ;ps;J*=5ϓ{. e~C<轗T$bB^r9.VT{/y~Y1 Pv7pE{agY, ZbU|W6~SߵQѰ}\R)r-e__kɾI_M%P]hokQ@2p!0}O5e|ً\ښ}6Z!QFm$ѧ2\|~p-~LNڭ8%kXN4R{!ijȮkOhrf*~'{{3 u5Ge-8O .IGƔP $F}1CuVazt^Щ];(L-+?7ìn6P5kN8\m{ې=)~u8Ma$kUGKha4-ST 5q#N0.:31X@O-rӎ;Q: ]l8o긽v1͹'Y->NP qM&NeFfu$؅TBk_QwTKNyZ}䎖2|c E*i*1~%fsysnʘU^PYIAѭV^ )XG}?0AtFkQxc9o =:39PU"]FtG&2Zʓ<^8K!&G\d}DM@, ˙„q:+tE: a1"NZᬅ0=)f4Y b0d]Ҫ5o-*fq`5IFh&"Th%>%)b@q^X(wCJ"?ƃHͱ A+ʨ 3A~ ^ɷ915VX* 0%$3$$1"OF!bWnu0'֮qڈѫn|*CF@ ݀S޵6r#E;ܢeXŇ9`o2{n 2/M6;֭_'3,_%mew/HfW"ʃPrlj%  o2 V8r4 tB~TB"F?د,ileA8?nOSۄtd<6QHAY'7ߞ -}VRo~psЅP|#jk鳿..8/ӗ@gBeo b~#jcg~F7Mڰs˳uY՟6:Ma[@o//y^ UwY:v9G#& 7oM*Dc8f z+g:#y۠`|%hK~9JxByt8іJdHj9緓 @t_o oozFoC\ċP]%q'CGv=ш,[$ cʼ X11Gպ( 5 'Bȶۂy"jCw8anށHf(Xʬcls;8&0Ld)JCDjkGڈ cP>gKςSXՇg?RD72EmN S/d+ۡdg͠@9)[_ c3@Q:`va2L +Z|戲Gs#;KUP(uE6HJXPN&_QqiPAk_1,W ᰃe_בzZ:6-NViucg[1mgvy1ܤ,R}쉨]6-۫ ΑmR$$h:I^{5Р&-jB)(A"۰ KI?7jhڮsjF} 'ͯV#a4X__OA_!ZW)nU[W)jw̺jUk2RQTF:Ax6lSL$WK^P;=Ǫ+*E UőyR7OWkݪoFWw8jIjHͫbJV%lq>/]R_9q>׻?"}mFB\0[;['wRaAus*JoW޲C#JJ/\fMCƐ"`*/t˫dT8:DbYuyI%kCDhM] GA"dֈZ^M VT~w?S;Tfտ3!NxBiN?ϒ -&YgN9!ͻ3U˗Ws87w+ /zṚfj%P PjMRC$/(y]CQ*ۧ UqC:e B*ǩK(fW}$4_Y)XseV!'皨bΊaʼ[]~eoKeVn?ֵ EfZpbPeVBr |LaU;ZΟ;0$;E_Dd^jZ7 drS*NFώ\͞V5[B(Sw'g"<)?{e}.3$x}{yٷ7>g . k=F +s6DD'Toj 8*ZtJ ضրcLKIAĹ#+@ɔ8t ќREqcS=۴(S(-JCU7 мdǥh&i_D'MZWI;s% [cmĩCYvTgiQBCF261 S;_=&I1%{$uEfYi%:`HP]L2#&JiEA,/54ʲTbgCqQEZЇA2-Sm7{73^\p2: ee6IxOM8:H2e&60QH钂 ix'&?ѠJ=McnK4K`mD+5e]\y.S3/p~dV23![eRqz-U `JjEϝOoqvu]}8ϸl>9>`'O,{ZpG'bVzU2`ۛTj]*}2MxYVXs^o](R>Sy8&b s >M̠MF#2[:pԸ9sbÙ)Xs0/3TWݥ =e{!2Wb@LE^kT"Pr`dv킙e f{}nR;ޜ@*EY*B+d)g%ِ#3}$h^ ܀|"d@whnmgu_/u-v iQeW Tg $ ~GmZT6 mm9hdxcJD[Y]D&rɣF&=u"4I AE8>Kw Q8Vvĭm\T8-?p8um/˼|YUA ϸ4-o ryu1 狴f>ieqR|T%+MM-J\:Q4f/Q8Nf_Uޟ yZj#wçc=%7O9on\^]CnKng8ڍחbEo=ɗ7V=L[kzӗ6A__YWXShA* ]PW1*zy1%n׷)n\9X%)_}+jݳ F7}ȄZd~ajSP]^*F&,׫~tOUJ4<A{:>p, ZsM6;.巏G\wj;N{7?=WsSce&a,Qfu.0eJg{qWVVfm{fM]~k߫!ؼ_Ճj%݌ oZA$tQ=o^~|,9_,C>^\e-!7S23BqHoZuz& 5 z$d߷'Pr$&bDdhP|w]q,ߓ+On`4fRWW;ӞIWmj8?Oͭ~x*! IVB¿a}WD#fh~íEt ~r*Z3ж5 ՜RiIIb0E&}(%q $. LLS.άR'_j[/<4m|%S Nf^b* M͓wDZ'.഍F{@d\RJJKK]epGI`wzR4A5?O'ɎP_N] H >Gdu vZ$m'ζHP:'㖋$ڵ$Y1Ѹ_bx'~vE8<.nF-n1 wj&hx5jwV&_ZXΝ Pi嶃d 5=GimGNǵGv;Cit\x|&iҚ;qDQ'p g.wgF}}yMe[!(l y#@ؖ@0TQTz̃%(`,dE*uBH*JB@.4Q`QlYJȡTN+_Cjsþ.4jMi3Py1%Gm$ *e+un. );^zazi6Ww`I[(DlUiN*K b%r,?I")P J56;tT2$1,(;\Z7sGNtr.j/IY" y=-`'ӭ{@iQP AI!cJ+_HaqC#]D;I%KO'zq1O.iA>CLͨ2Ioʏd!ܘ^?ey/)?mS,S!}=P~_ӑ hhՍ%xЛn߰aYEx{i8fhFʗkA+^ +43̐"fB!d{qw*d7SBAu&dԔs\+.uF9G;Ft MY\xt'YsdY_竳G_o}P=^ b.fquGnVd.A}PRgSn%Q >^[%[+FAKEHrhԯv֐VS恾| ШƑ'i?tpXK^CTݞhޓW`9a/ $b7Kl6Dz5YIO7uQ8j5"c]]U]wUXծ"eD[ -J@ʢӘ&#s/ ]^~PatC .ywG;bbbI!h3^W5L_~08fB Pq5Y-a(U}xT%Ҕjh˄@PLRy(/(H& Q' Hx 2AŒ(ㄲcD/_s^< I/ⅻOh9E{n~~O勺 Y (C-'@(V+lRxЅnܴ-vDZZ#! $E^e݄>~ϘWtYf`M>9b *?7W^ ݄ u Q CupJuc@+נz^]p/B]<}պ`Bڵk_"dPp|%nVYḻ)li^ &6f0b"[JUrqc橖koMmlzP6ahlxl~X(}rVmt_|"xAollOzv?^X &DNoA!N1 o=Bٸ,m@pAff ν,'u};g`s,{HUI*|Ttq[Ha}à:M˦7]{Ʈr+6r " =K^јxMv&S. NI-/$ e5uZ Ã򠤍7Xrq4W`qVjv1.,(CYar=G.Mz5|8!|<$f3 O3w!3MD&#"c%&HCf68XEE\rWDxCaMi'5;OwjQB @p tL+ksXV|/tO8ѥvX VǷfuJ:qJ"Dy}b݉,#E SzyoN P]x ReByND%=_ĸcMQ}@}`F"R@/o^p׀R\#炐mtaШ8ch1y9 V~Hzlzz.Cyoxr(o~iayw.Q󫿏YxPr`Q>4bc9W̮/bt"؆AlzvF;61`9mxnޕC[ׁ'626\5[G;Q`w*_^sȂb;7r˂mLӸ1'LojUM1' wlyO2* Y=뻳7&;dz/Tf{,>o;2wyF$5;ߩA-7ITgm0`jw^1!CW}: KC X۫M1o~Y=-wǸ)i,bf|ٸ@ i) 4c*^AZ89OlHt$[!Bw_<_M?Lj8t$$0QFk)Y9((FC G u՗`(`'s 2T#(_PZcMc[%CKdC[lΓiƸ=}%5pk6~ VݭCϠ]8[5U.:myV"r{^u7X EW$ 0ʦ0(U#Q'IfEencD`á(= 5D0 C$`T҈ zswu~FbF9?=/O& Y",A2#a0J ($5J:5 sYeҘ4ToHAλS6}e"`Ah %Dy5707e}yQ h^5%iRp~^,4@۩aM`kLIX[78oKЋ ܹhQc(:驳C bZέy#i=.?nNG{n<a($aB 0coP=?ҪD%R"|u缔'(F48"H$qaD4I3*Xs^)p:9-K;Qڪ40_ǤdVuΣ*?o=|\OcC_,[;V@2GփOd QO똷xny8w=3DN_l;3{9_.)$Lw;DާР{ԝ#(њ=˃t3Z\+hpT.G.l4$hbz?{t& (xk(.={tqArFD֡yT?_řOa, iOƍWf[ ?KJ ^V=!7䜩um*pӕ yĸ m8+{lu`5gj|l>&+*JFLCOBYUQĐ%=|>=9!Շ̚J p:9X533 ׶GؾI"ZtIo{jGĠ_7`-I\MqDyv<ƑO/XV1I\>M }{:F.er7UsɁHp9E&eJfBZBQWTޚ۩/nv{4'rTNWt' 'K=fH~V T5HT ; =/ NQ68;?%|Wa]!=w`Sܟ滹eǔk<4۽c.iي *PpYa4MAjp<;m6{ B@X޳- PdOf%=J=c,HRj2JcWr`n8k AkXAwFzVZ@N꾧Tj>?/,]|e7%?|ʺ!L)v'D*5V UFWc 2X(ъ 5_At/e(Sp&A8pZ++E2`?VT:@IT(H q^YP ?([5'2MXd&v5p#:](FE}܇4Ɨ/߃1֚0õk+U|?V\b9C_a -SFz] ƅXV$!s͑)r:T[v3{nNn]Ӕ+=i쁦j>$o.tE.TD!(H(]R*qE4̉'7Qk9V6ې̏5lFRl`=\bTHBzɕ6_iFj>I!cn;x׽%2Cfp^^z}(qMf8>|/VOOn)z)Uuż3aiBXfLpbgţzM v@cw B=Cl̀Mz3e paDɇ'tjclkZ̳X|7M d c0HwmZRד?h8nFѦz2Qpn}ђѦz2(GXPSy);s|H)L13mN3Nt?sHGHk)$Aow SL[3.e3Mi=pGi<8;u *bEiϓzqS'DO/e3ߔyuΣ v KqUU8HI+jEl"x cg875, o,%+r;:-Elb+%Nb~@ewJtVxT6z%mE9qV@VN c%U#ioJ!"6iədFB?Ϛ+ns3D5 Dt ejϺS/m2G38Ժh}pJ^l=Ѓx͚Iڧ_k. sJH~J\)KV齶z7`[g|8<%vJ(?6l <  GF;Z+Us=mֺӲHO ]$&|xaRL{q +F\W^|CßAzoc#0N~x `'Qє`)ȴh}7Ȑ23|[HknP%!x, ՇWIK[qSkyxHU1[wiFz{~:Ґ3.c8_4u~9/]nIxxLj! Olvi0x6G mv>6Oum>y9div öxxc;6.D&j%ǺZb}7:ui(zJBqyMhZP6ڱ;i)lMxrDݿz$?ZV{MgS1Gd׃{3#\d0nW,OFp^`B;=DHyt<3R sZlziBju5I K gW!mdO? ŧxr^_Ii;C2 mtqI&'']Uv@WZIgzx9DT&Q'J EgP0Tt5$;N1r*D: 'xPNW۱tYSRe&cb\C )[V$g5ihB*q3ZerVIԘ֤0+TfjI97.zÿX"].fY ߥl(*a6lavQ~}s믋ITo_~ *P0ƮҙI"kɼ%HYcy jif o>aL˨O&B?:fx-'ɳbfs|46X,' z)mP\<ϐLV(l%OO #J{sX^Ȓe:=Arsif% XH R̀n0U%+%[=^4^̌ ./1[f1 ќֱؐl@J!K+)(0RA 9Q j7i=2(c5NgJ^:jq2Ik(5NU~ Ih\vE%Ŧ'$W5M"| yy[!æT7LB~k &װvo /ܘ &?^ұݲ? (( XW/_b"ïy0}j/dn'-^mRfEn'r (ns@VT h6%<cT F eDDXpHy|k|$cMdq!Xdllr]F2ĄڎvWň,P|koʕKcU Th80l3)R ڱep!W*E)F)SP,iXb ]*O}wc]WSl|[ ϓ/./᪎QgZ o\֟4O3Х1uEMnh%QB@d j?-Z Ԛ A45r=_SNo> `r^HL4!OӰׯfɋq'{6 :ݝ-WzvX@F)ٺ |>]וn/Ċ?OQN!油5B\" 5)6ZFQ7(w^jyūY HV? }P1-ͬ#VcԜE`F֓%z?1bRe~lDԁw|CQɆQ7zQOaTQv-Z~Hγ k̅=rR8Qf&3/fwԯzBr:yH^~TDu l{&pѵfzҰΟ}y!R'>MbJU17Z% ȍ::M4|$$g d% qUX d;L ÄYr+ma0Kĝ0XIbGıxD; c9%Jмa/%?tqF{b5*T hv']_WhT*q _j/@Rכ@g2Y t/U/+MJJμCH92ޔ%n=z'L'Ji=vfzuyX"XXϣj0y7k?]_ó exy Ma% I3<8,kݸ=dٮ)=1 NqxOvyj75[D9DLX6=TN2 lJɷ-a?T'|t5Ψ> { kOq=AJבϤK)%[!%ǚEc%] RJ3Yaw%@ cQ ?OO9EL wޑ|j=8@ Ny:{SK)ݵCWYzO#߾]WUBb}ݪIʅF|:4U(UVecݧB0H,~g+E~T\ mqbF/gW| 7?1,&f>t}#>]r'or7}5x}3ZrVd D􏗸 Ӟs+ԍyE&Sq K[5lZ^[ۯ$SDҁw|+3gw /7h":\MNnfG 'F3/^zyya__2I^wg QՃ~VBՃ>Mȁc=f[5># 6iiFP66(!P绶j'ZR1LiQʹg+ ܹFzҰu=[JyQ*f}^=ߨRvʶ^ ƙU]"xf v'Y`g Vc, D&(c ;cVzd$GVR%H : QZyrեeU]}y|׎|hM=ePIz_վLQET"%bEATYi5IU lz4+%v,aP6%{r({DɔkX|QwAMK6ꪎU :ﱫ!>dM|5 Yi锂N$m0$QT $j7%'N;al,"M`KHf8L&5*#E#L㤨>HH@d?P[G TMs ;M Ո39kkXκNՄ3ـIZ-D `IIITSrohƄ$OF)}靟7>d.cpqIUKG|!m F(`%wxa" dUYQD 3Pq#d,2H4@wh\Px[윜L$owJDX4H%wRC&}] W(鋽((օQ.HGEb2B*ߔ' !FNEG&r|0ܠ0Jz$~$ƪ0U73뮙n'WnKEW-1oa) 4A-sݳf//%; Q'CZVF3tWfzPw z]Γ!WH]U]=7iž,Ѥ&@22V]~2FZYxHy5Q<(038S1crϼY!`fp؆т+vCGuݭE#؃oDq?Fo5Fڪďv=x80 7~!7QSAcb]aeE𖣍\G[ay !+EfZ!cr/3i<4vl(3&s $3|-€g'sH-"ęR;RcQ8c _.Z 00(Ò@0بT,"zφOIT:&u#XIDB<3ILDH K-sԄ$XF*N%z ^G ܧ:=ᴾOLN7&N;%Ь7:s|y\zn8&0D 6q=VډBl (]9`T'k.#ժuAlɭbLŇ:v~?#h|!V~}z$>A-=֬zqENp*PU}_?q7bO1^MO{?nΥiO[;7OXp}MfTAʓd8r-/sO_:qK\AZzHټ W RFkP[ȌYk-4htk㊗Ϲ6A$֨aG"Υ)ߥɗJSO@I_oJ{OO$S22%;cRkJLb]um%,b~6\nڨ98c\ $8sPt<5#w\FNÃ9K&RȲ.D&Ҷ JF%5}KBSh DR^22hAj!@_^aD @ zW?uLq.D"1UcI01Ȇc6)e`dpm 3`\DGzI^H/AؔFxi5# T,!eh9I%G,2b2b;8Po pV-S=-S=^&Xhx W8ŭ fOdԚGQ{Q K-ew`ZB" !gL[i-^[){E2DEGy[T3h:& 6R.kP=XDPYBf:6dez}}%n@E[IUG]_dc4s2 ctyyzgYKkpol`H|bltLQ[EJ hB}@xY.t3]ణ}'1['agir2P> :.鈕eu&8 Ɇ Ҫ05({r6kH(Rv E}pkMp}g(u %{KSĵjmq}p& ˆC(L]Rku4 i\:7KoSOUG=ʪ^Lx ҝ|ENh'u[ʟ]ozAQ_ݼY~Ng#mT9$D-Jk+iDiԂCC<;+MiF6ׄ7=>دI&,V:N{jMq9?!|)E)zE--if8r߆GwșH" U3*Ld V"D@'uqgC_^+wXMXRRsXA֬|sմZnYUj=(f1b M"nEy5k8lFјfGG5G4K[7uոس63oŢ1 { S5kqt6 ojtuflR[Nfm ޔgd=i="˘=i=!4n< -H)1Z4ytE01uYjߔ |;p Kyn{ @7^j( a{6F*9{z~|få70gduSrT1ע[h<>:uFe?ґYycsma0X#w?X ,4ׇz-{%t@( o>>wwX_k>8=d츹/ ɲ3P LWNj#c}_2~.'Hv.I:dě4^Y&;ݽtVHdf {=>)K}ߗs%t˪1Y%{/4jbPpt1r`088+hqaRh80Esnٖۅ΍㶧hTMΧEM3 pyyͲ3OݯJ>AK$Oݰ>݆Y[:CɿiZ&I8+cSz3;}Xm_~yt/ܽٸy}+(x'OJ:c`KڧǵKJ;/~BՓͶY*mH4VRL9RT|6ȉ~^ӛ\.90SSlbFoYdv<P gx66w\ְHTG?y5_q2N;mbvi6nOƢ tbߑ(hR$_DjZ(\rkZ!56307߇RYk 1Zes^Kt@Q;M$Z HdKGp1Er)"s{Vqmnm2Y,U``)V_{wH T\H0(XWJ8NR"bR Gȴ&HSޱJg薼4BHPJ qp Kr#`5`SFl.yN ӈt@ )$;5rBN"3[Hbm(tҎ(CP2nl߿0 c0 jnBPТ,`pNĎ #PXDQ0.-h~j)qe.266b26b?YPR :M5 _wLpĀC}ctt;K9)LJ [R)K : *\%EEB k*H;PԺd!Ђw',B\_0Z~ckWO]h!PQFa.4I#ɅԅYH_.0(ӻ kE,mRT-uãe$O \? Ko{h.N},] sJRAJ.8*aT1_Qҗ̺( ,~=ro FnBm)$:a2_HpG˝PHRG%+m(+ʔJHnl-hD#D JpQ ol%31GBip Qj52m e_ F#,WFA48-HxF'q Å+4gsQ" .L;*Ac} 8QDDN(HɎ$wL$¸aږvYe v`fXUNt\9AzcB8C-)%"` [-|+y B ةn5xWq@5L fhJ< Fp^Tad6h?:؁/:z\o2Od2A N4OQd6BoB.a&sQ$tR|&kّ>KwM!?t#e8 p,ՊA3\= _ GۉF@[ǝͻ Dl8֘yNopI;u 1k:Hَ1H Б-*EgJ< tH%2<+<6&Վ  (D=B_~#Ik< o?-|={Wyt68fQ{ߧ]j(q2]|_1OFP# OQ0v B;d}6faXنAwb&bI}s J}-TzNb6̃b\$FC Rf<| Kٱ ~,~1d,( 򂔲AHxcR.D,T%)Ywʓlʰ˼c<#,Xg/luJ4MGә]Ys^ϓ½˗7GhpzH i-Km%9N)Chrp:Mo4'=Z>7Zh6!99UH$KU:Ͷo:9/MHs "$!!Õӝsd:\iQ%Tw==pp4P聣TbaJi<|s= C8#AƧxxcsmg'f4yӞ7y390?ȴǝ3k3nL¾ OP^ TpG *mN_&pT[O%iL.>Z䪳!Rj%]VlBҽ!uG8V *OE9c`䃈.B1Lx[&3zvJ:Ʋj,q*)%E܂Y/1-Dn`s-bBruE?Fo][+ymٝ>9-.|l21Z$o|jC6>zvӈ!dmD{0$q55gw("|ZoD~ ad>!ѳF*w܄x&V0h-)9m BIJхq߈P2.n`ZG 5tpEiMF7Zbt OcߓzZ:^MOhQ]GGH*V}Kr0'WbE0E\~9)"C4{jFݣ?bq}Z,&w={L#%&W| 1z͹@JbǯJ# 6t {J|^*'H_n}Ԕ D:eszraiu 9DҼTR*5haE9R\|%YqIbT#&uig595~ H'Z;‚ fB0/|?_a+sS8dD!SGA" SdJ*)d*|"SaLHjګ[OALaoVypd'˫ۉyyߋػ$'/AvIVw'"V5@8=J|LAS>T)uߦ?k$Qwu7p;a_n"/Uq ak-й3ylV䙁u{B#-w[p*~zo<&#յphjQ܂x&;wYCs\\#DjmHf4xҴ_7Ɂ0H'ϜleYm!6DA"DstQ"( W]aAWsα#!2Ԓ0ju.Wh!HFj9o*rvQ먵oN`^w_7 ,xYfE9{1gh|C#Rз4˧1:~-A_&ӏOBl.ÄSo ߹\x{XL>N@7T'`JX5 IZ 9ԑi*,f@&`QX ͙bhjEjݱ^'T.]aG:P҆4F4IEa`Qi,i*5,L)=bRg%"%wfWq4CcWQװϊt} Jx6?'W3uY Lu=^`/ :Xs`ex\|ll9;VU@U6nι%uqA4mݘĜ/I8 uQ85T} [ %Jp3O~M3YZ#gfjʈL4i.6s ݧխ?ĽH^8JOriSH{RM *6'`u4e6.Fv9d]p9[#:'OS1XCY!"6˽e@˝Xgph0!ŵLHz&Қ2v}< Na`] % ǁ1ΗA.UG'A5ܬks[ \Oo|>0ѳ_ dHV{*'*Vc (ZG$-|$%psh+ǁh-ܔDOY[f2S=4+BEb`#h GZ9|ٛnP mJP_HƪUhAB*r}?5OxS<DjԬyEmmRkVҼM2G$kIy8G6܀mż9`]k 5 iB!Kaa5wVH 2S1p2C΄0 VѬGjl!n Ȇ#[)xY82yr3,IXbj CBeHOՌ#oRӛp;+h |h5Q-sZȤj8*!,;x w6xq3[,|y}v_~vj S f|3S!7,z9s Q AJWѐ ]gfw g pE\0h_8눑şroB&!('Zg".`llcE4,.´в ZK.p ,F+&d+Ji&;wm~9'Yފd 0;  ӁмޱYKs TcO˖dJ-uK,k[jVEU /!ZcHik-hd],f`K5.yFб͙($)k}[ڱ \x_Wy曌xʦBd NPzKɰX3\LHrQCbߒG[$%螐xâ-KX(^&dH0Rա'1*rxѢpiEvUHe$->Z)~kY~ہs`p7XficCU>wPTUwNHjz.ko>U\崂W12,'nv>mfxeL0y:3їޑ %p!84[o6F (o)Yrj]&N珛gɹNݥ3=ﻨ5Ypc[*8eq+ π>T&rQy XpDKP؍j R 猰 ͱ&,f{Q2 Pȑ i'Θ0|Q.{}-fm k䮠KOVxKӧ.0IpB,Ȑ_8D#wl=58 k ִ|CN ‰Aڿ:W^B+ց)goLpRhqvb<]W4}v֮5ζt3ǁ5sZwV~<G)o,9V RA`(T@QXbP'DeflXv5 匴RCRa{KW>_y1G[;R"YpxlԵbX/S ݻ'`e_/]\h:Q(iBB&X'q5ܚzľua}e;| ͐(٪g]nճoB2iʲTa-YTdcI}hz]t,44E|l.9 Y3PcU4j$N$jP(옅4ta4]WECu`897NմӅVֲP{=~wمTKTUB"5}qeo2"8 RX[(WXDST7}ȍh I.wWȡ-6,=Tw4"yw rsO0ڸ^ Y ?(쬕,4d]sLxEi:b2 Rja6v"PSj9G9߾vTCJOmǯNcno9]|}#cwTNRӎG,-.b.ȏU B)+EY(k[KmۿFU˦rTVdTŅ896<؍A) .-*QmY+eӏAC|zRMeH:+VC[~s@>"ImKAbewgnYAi+dc ~#J`onӟ^._! m )VdyYyYm]IX[E~.ŧȶ9 ָ">FSŋ^`,mtzs3R>K75R'Fݛb r Dzǿ. up~SNi/@\$tA "߰r>sy?EZ~W뗯kv@jЊhjF:>X8DzKA?{,DY&0Ɲܷ'n ,J'qDbҰedH/=L4dЂi-1dz_yyMt5 WӋ3Z]wgMz})ןb4_Nkl~k~"cxDκ?MA+BC_AΊE񫵨+&#jpk  Tw?R;CvACˆ7B!`s7'/`k_ضٻrɦ;i78VE`mOTɤsM-6=/bL"giN> >42l=q2ׂAڀd"}!|BvZcJˮJ.^ӻ7nƷy&ڢ{=}.AIW$=-t# L":0ъ~Ч|`,|*>g!"F!˜@$#mJDB0 OxҀeTt ) g.rNa ЩZI3Ky2ײ "!aEϹ ,E9ߛXgOQSTA)czuP%,& ä*;_'4y SF~3b&PF,$\wۇ>0[<ǻ/W7F@L/Ϭj_~xd8!"Nm'N%(L}#&Bipk@{JcɂQ1mЇܰtF+;_#_W޹+z |Kft(d6j`HKdҵ˯_b)7,I=dx&61ِ6W$'q G' ! mȫ TAXӁ->; ԪE k)5YlOutJ(2%@0p"\J^⑓*9⑓*9#̕`\_z _N34l,5KLNC%~#c9Mjf< ~XH:ɳ}`L Ƥ`L ƤcQ[MӾ$QK+ц'iRE&JhL@qw Ygw 5.iC\B* e-jVJ/}I睬bd"=@9/ jEK A]r2B[%i~wI+e*_p$JX:g& dmNJ5W=V?̟5r{|9=9i0"W\0.}-Y:% H>v^+bu }!Zr: j8JF[XBs0{l}y t*!*1Kn=:K#2M&]zu:Lc͔Weޔt>!u_o혶\.CN tTfa*HjвU㶝_/hBp֋µG!WX[{2mV=aѨ9yo2y;'VA iYqUъ.<nD=Xkx&b2|mI_@3ƥ.ɼ'A,3O(_,?(u vZ"02BȹdN< -04*1T$ rMT.ѲOPb ς-<| |HQo d@m˭j4bdWi5y)Lsikξ[<}W嚇q0 Fi O{򋹤 ?=Ffٻ&m,WTٍU0e{]l\$OrQK% RH ^"f춚"8juI!og#}iAY—_jw> [2/^4l р+VM~x6U>>5(x ˴Q^ڗRLS3L`SW)e*i!EC, A3Tۅէglhx}Kx3=,Xt35.Vh6s~[l2{8p؈`*:u1pA3{5Kva-QYSƧClHX4IB@ U $j Tp,BNQ9vLzHHHҔeI42XE!z<8eQ1 % D8Y c9HA2!"ɠqu#Ʃ$$a(n'z'%nO|ب$U=[Fz՗u餂 dYre~Ͽ+&LPn.\*՜`]~ %ADYƯlÂoݮW"4ҳGUr5 B;G/>'c+ps2ݮm~(wKj b8:(z{>gi U=JIL?L]9H'XKl7֗c}8]z63ٙ6LV3>c / la=ڻƖYJZq֭A;=M)2kcO`pc l)e7ٶS3{Mk'$[d;T|( w;)\M|r5rŴH*.d?QfA!ulVc'bUca$DE 4?E$APK=B%Fr&SG0,AtqZRg$\aqۣ޲.si(GĔu SDZy ZB\Ela;&㸲O4ʬ+ΈV;#vP)$Z!\F!pu_m_T֞إQ:4m^Wq^Q8[< mx{ĄBne,V^vF`:F.զfU{:Y,p4k+UnR؊WKސB#Wts‚re]1+Ιb깲f{iV=5Q] МJS@ث`7O?v'OG}w@DZ0 9z*d ,444#DU A vKCI}CN\@e#BKmxJR_~ӓA+KySsTwYf P {b!'t[UsՆWL$n8uϮ3Hklb#8²{?޼imM#NXJG61=yȋOé^֤7V-'{w [KOǫ(N}.clSߘy~PQ\جK@=Y 2gW&/ #JVC`ґ/ "p&B9'-gAhs]~Dvkgy>0%|XejBDRו AG: ΅6:8+kG!=ޅb݅vTP*Eޙ4xsTLZM >q4 ؉{swcqAԦR &sSnہ՚A-E)Gk-PfSWJRuf=9_2mfjȔK1s* ㈣<6?YGVe(QX;=M=7VƇ'ɹ|tͫÂ]?ɖ?pm#tyy2|DN+A,٧ĬX{Idg*z$Ɉ_[uͱ3{ldCɬzmQ̆_=#ʇӘtX\+6Cǵ*ڭ=)i7:ڭrCiLhzGas&h Դ0 W&KPi&8\.5 UOo>96nCZ$ÐW{,ӹ*,`.49G\.5fyM13b6_*#tR"/^8'[riq^GSi dVJ$@$ ZLFqA(;L Qrk12,!ɚcГ],7{oy+JPnr2UZ΢$}6.NeEͩ}f7Yr8\$seZ50˫A#5\ۨr5 ?sʚ3'41B|)oC:n \1:mTn- A9R1B|:)ۑԔP0:_L$ݓ"BER1 Ú>GxuP.S@] 9Z($RޝWEv_kƃ r̐ȻgODU ;9眠(T@1P䉀:Dp5rxA'9UaTԾqGH#,!~IҀ$ C $BРrTٴeCU]TyW@ރ8K+b*%z]޳)W/̖aELQ޳#j^+RNDd"?\fݩ$!HˎMd5ݿ_y}{pM wU*  Jk&x-ܛ `X^>wV}ٮj/[ .v.Kg}H&³Htzmڝ=o5 (j8|gG`n!x4A@ܸe{ѵ\@b&} Mh^l:ypOUݗpq7GwL;7mv<yzǧ:oZ:q(@"L P1`zzj@2WF|zb;.`}")O&=tq$ TX Ҁ&ZQJA0 !UR*B(UJQʱHaI""$\$L""T0oY+|.a\ /jڪ#TlZt\7ʲYr9_ٛ;wMJq$"2ePa&_Rxy 8dsI8T dDB!^ ILVI&a Q*5*'KT#Ϋ&-ή M{F,wj`y{9+vԯ 9眧 N&'o6vM z cן 6: 9Hz,Gu!HsHW3Y{d($B MXzBR=|u1F>rv|^alg37{PśiQiY*8*^T7[T3~q;q:=Lx{moI?J^f-^JSmIYGkfHeR4T٢wgPCB;qGX)4>[s{v7͖ Xُ'*kcC>J9d 7}I r6yyEЪ+8X0DmyńV(}D4j/"ٜ+|-Φ`\&igrTqcO k‰c|b5Ż_[7wf~9 ݥ\3S:K}zTk~PQ\zv(|-pfNu V:CĦX0o5TM dSm^ШyEDAC@Z~t ޾񛳅۴QaHdpnI|Md~dZgŌ߾ƉЊ6zKI-VEx'J> ȑFH5êViU:+6nghJAފrr|w.jg&"-)< g<D|qQaͦRʔ"XILCGr;ED\J`s'ڪ$0 O 8VFN+Y*֜5 Uw-hOUIJٽ?rVD\sp-?~l%%ݗpq7GwۥU嶻Mۗw^E(%( ^#B mY0ZnY׭zZf}?b,.;ט喀M9J^HhKPsL &Ǐsݮ̭x2+6:^}֓ajϒ< )iJ2&DiFY4 1ɣbbW~g^7]bP'"TX!"TeTOm% )ac&u!8@Z" oʘ/ͺ^Z`Su]zūx:^~V@e6^% V⽹A}ѫV!WXu㛛Wv\_o{3ְmckAzo{M{E@>ub#vM`3CK,)A;CR%[2%.ߨ@cr<,rNsit('tPB$2.n8>x]{oyuy|g/'ȏcs[5|wEw:ٳNKbO~U7O2}$o߱nȁ6.zݯF ~T8ieąL}Fضu>8i`3b0DV:-s{ yۛnPt:F}U :86 t4=un4w;kEoE5qpRuFN;~ZAul){Xn-u8}oKs9|ۭ g%7/D$WK!w7k42I=?vYE&xo=mDULNNO?GYilUn`Ӆxxf}?YN*oz}}W6ݸQup,!1IYJhO[9 tadfi #g4TC=| _Cr$D /0bCBp}.V%P s*}Bi|f8W0(dF|q-`ƞRx yb.[Z,"̹"bEso8ΦmSKښ 2˜<3HàcGu YO'SdeWOL_M Ac<׋kg^7giN C]xj:G6}l[;OLrhLB{`okR\;v{7ɻ8xw_n_2/ULi 'zV` ) ujm\ĴJ09kbd;{{H;K)_Xa밀z$_:UB<{ky1ۛBTx w ҊRl0dqkrkc%F$P_߳B7ץMmA?N^_t >/֖l3$! '¸ ƄT$*FRX2+s)I+!I.L 4ʷ %9A_@4JQ}xLsj\edQ+M6wOf*uG HmFOVJs55c55TVˆa-$oτC,b.:XHU|֩ evlp%qncf GaB*twOx3K~gy*$ɉ<ȜϗF!!@A͂ŏO@r|ɋqY6N>w‘7lWၒ~0]^hC]*•I՛|qh2K^b]|{;bH|?*:\՝t5 P~]-Ki+T#ukV>x'A9ϥ);E5Pa-I]Na]TcѲJG w^NC.X]suzžM}.WA)珺Fo{xmisa? F%L,|)|x-,t~.#Iw*Mڻk@mBM<8.O!L8)VT tLpHQcda`<Qi37-{m*1}~NSAH9D-*GZOFT 5a@QiJ.a .JLwK*jk_!q^}铼{sR̤J{{#f=%kw} mi7Թ:{#k{-gոwCԵ%BZ 7ZJDlope:Ä(,V `4.V?Wh@TZKa2۶\7buxzv~v}l$C5m=EhbRf|}ϭ)$<4:, D0\JD:yu,]tf;[^-t\0Cx9;Q&ѝ=b¾^8!,Q&=Eڀk)){7.\^{SK<`\pkc0neVE0_qR1sVעH@#d(}FKM1<<&OOיdڜ,dJzwיv8[S>Q Y@0Ȯgdl#twYMb!,BȮ"cق[{ŵrH-͕5("ܧe@I3 :B<0 PϴbdAYLI+x56fdޑVmwaI1HV*ʏxo\t i,#( b2;-}Tf cjhMx3<%UDW)Bq.G?- fPI%2\]ev#yS,5Y?2ɿI$CKxDq(#Љ\4\tI´f9qlKu*Gèud1~2E43kЀڻàCaUΗ]@j*pvGmصKO3*Y9?|WBI8Ǜ f8\h   0i?1T@ |-ϴ-$|+ޠ?TÝb I mE,@ϴwz"͵PF\KUX$G{ Ha7f R͐~TmJpA+w x_ey_e/}9|EOvS-Jݸ\=9h:ET,RzkM#qu~1y9D01`gtU Idhgם;nb^ZPТ?`k>i6kqܸE藅vyH2`ذ~be=N_n5["HOFPdwN[K\?<>!uc[|l \" $c˗_/{jW|IB\,\ݬ~nߺ=A/L4E6`bqwnn]ic [O {B3U Ɨn1IvARE.įW)jɀ4V܃A(sTS- J^HdRQ GO%*g6 EUtT%q-r=#sT<|JJ tJ}c 7"JE4 Lr Icmʹdj}wW(?m _{uqe_-ob\7vw.3-V{>ގwR j)$΁YrEɲ"]H J x՜W\KQDQO-* @>*@C 6rXQQQ -u`"AX@3XdڦB|YhZIbƔ':K(]*IT5A8z⮪*dUٹ֒w M_Kf|kOt!_[XYoúG~E20%ۋ,Ap)BqbEni dž  E#iMր9g$R-S{"s`0A-DeTVYg dKug}#ZyJE]R.FC A=nHN(j+Ttt/@6(:.M524 b >EJ@m]x@{ls_> >=}?j@.)Kr }%`_-V_.R > ]Ήy"){wma}{Z [eBHQ}"'NI}W+m],Uj5Rdb0`VOwgb=_eBDCݣ/cenݻo Y+#D01Y$9Zb>32c/"M7K g749볡#y*AYK1 -(r|haDdʫ9O :zj>3EDN%›V&6x H35yFjR93Ҍ-rq :g0`PkzuzpC }:?ӆQBC'DQ*}76|tT~fQ& }5{s2:%DnjjJ .*n^`AI׵U7p*̼K Z8J Ij9DsE]k-esPX~%aݍE~׋^dsvGu1UhG x( '\TS4,_TC:zru!pT3<4a0`? /m'JrKB v fUnssJ;*d Ԙy:jEsE'CvhXKVQ"TRS o XtDH92!XUI zȂ~՘JR f': WTI ߴ[\J(%:6f5 ^^zVT$pRGI>сĴǀ=r /UBjvT=h [#~,mn)BжA/R2i֑&y4^w=/86$'ֹwm\Di+x@_>n{xH;ҩxw3PQ";^R|cҋbBF >apZ@~ ; tvIg5 P; !hA.5|2q]vϬƽrM,T[$)RE:Snzݔy~)鰆?yR(v;KT# LcY7ynߗ~ue9u/w4ǿX\@ ٩9N$b87W%GO?P} D/q'O ۏ=+' wOt)cPG_9Z{0u>C-Q-ߘ2¡MiFS4Ai.Q)W`7羿 nC|33A3/.us٧^.qu-^b㿭'|DF$a vZɱq!x2 LOS+ 8D-*RT4L܏޴Ҳ ٰLcqL ?>lZ.!y{k5Eo+^G3`,3@8G`rPLv򐛶YK[h!NXg܅lBĀ!AI/yև {Ewp(O$ߏȴjGG kYhjKg9Hf]M[Qu0K_9=RP,vL&LuZ`q5 ZqNSt3q@|i6KL۫RerHKBIk 60\`1Z"tFCQJO$x%3qtް̔w]-lWwsUgI_)ޘxԨ}9t)}ߐԻ^6hgNʌY͸++E(/mL+A-@lE JXƬcWD";QĿti&bX C.IEL:_1IXqM!xL cE+m*ִ1߈!qoe!([v #KJ z&ձћ]֫ho2=:bFL8=ꐝ4yztݼ&` ZW9P^gT$ dxwDHsƓ@gʻC &!Ev2˶;tsg iWQeϰ A"pf,1*͝ь B͝5^S4MV7@(c ueZ~ҹZ7q~BA_GӇۚ9mY*#J'&09?nQ}>\͡L! <e7 󬿁IYEĨhF1@Z3#-""g$E7&D, T)!dmx@ʞGZB*L&Qf\u0(B.ESղZil&FLimtl}'#;ܝ*"pW??nRKEZ tޖvi#2i >( r7ZN{י:cH== I쳤ʉqr/,j?OH2I910gΓ6JGn)YK% ĿlO"K7M$3ubqgIjnβàmyBj` ; uFnе8nel#hWM*MPfe 1dF.0AxsHPEa:,l" 7P+kf0ufD񡋢j$ƇHi[p/i$l`9mk/m)B*_:PŀM$"9H>^ZcqhT4JX՛ga&%)W/fѣ./MdEYx/! as(X<#)GScrh0j$pƇ;Tl䐧RQ{DVMsufKJt,:)BcT-ZCpJcEq3Fz\Q_}1&FN!^@Nf.[^+m#Ilcȼ>؞`{FԀiRܞ")3uhpm1YEfdDdAˁ\N"luYc3|Tlt5V$)Yq9'hL;prSֵo̗R\|Vz9Ȧ:)8E]K}ԑaIj]&s9m L鮢8#Prڌ :BMbaX"nk;&(ޙ2J0V9h`$[xN;^|tKRjG'{R5痂R?.C Sk=,[wٰo9ǝT9y*pL 8rȂO7>Ἴkt~H~y>@S̘če8./-g*awײ Axׂ.*0g$-F nK>_"Jpۍ q:k~AGIo. m*0o}q>1&oWw;fDyaF({0GoV ~P)Ny^9PT N3MEj<# BH]pK|mܜ|ʀ?ק>~ݻ6Ľ87)͛#d%gP1lp$41NAX5]D3(V3AJyk$Ÿ{Ph{EN9AO%?78{ |2{?YTC]\!U'#oywa\ݷ :HP i49gWw?y=߫ˁ=0@$Z.hD\,:c$f @J]n_Of2|Ô' W+Tڛ|vM^&N^P1ĩ*I(\(N kiWc~<_ʔh-煻EZrE>(R#Qm( 9ehlD(kҲ |r[7|kl~ _^8*b:RI~#s$<0</0 | 0=kR`9 B d R!-L+59t4!1k+bc|`HTX)1(1>ZwL M(gIE=c@qv+7H  QDF@Z*-F"d W\$>ύLJy=c{ Y4Lrt?.,Rh9?HN#O_FADdfŚE/`M%>=ގ׹ip6Cw;f:ҜJc6`ɩ%c/]#=Pk81+Ԁ ƥ]insQw[J4k rWIn\AvSzjLv[֜,Xڛ2Q0 2URxEUp B`vK0_orSgax-8,!N%X6A: znGŧx#/sêpNJM\f' n*5sTQYIKU94+QvyN]Ӿ.mQ -v[{'cZVs)JĭSϟΥ!!ݒ~:!1pYZW?O`5v&`V &m8E`(4ltW};PTp,~F2 MK2r4k=2mՔ#eQDe9)OtyD\kriq^ YcJ9*g`4<`0 BjBt~ s<"oe;-vLo4οKt-*Ml匼Z{UyE?r=Nj땣y8A&:j89"1([J)1.9FT\͌!:dVH-wW&{F E`ֺwfl1UƟSpo$h|" ^BuKUSW$^0>߅EOx!g(?s5`:]swk OyC9›ݩ aT}"U:>[6TcsWh[VU_;/$]L[TȜk1qϼBQ"%P/H# dXi;3j)Sg3ZQ$72J'< Lp2bWP% mPX)ɬ< FuԜS-J*ijX^&%x{€o^&pB)8j%XY)ez㖇6UaDg};n_=9YS&wՀof(L@%P\oaT2M*z^pPbmPz(PPpi*~28S<:Gг* L1. `ng7_[ hfOD4U%D`l9ղ ]sM IR,TE+#V[P7;e-9 gzg LFH.Sk:u1PXQ,7F Ƨyg#s6X 4z'epӦ6wIV<72y]WJ Bӊ瑒';M%5*.,XqC)FRL+M+{^*dcLL1TENgw.'\sՎUsՊx J r/ Kd66yπ/#"r+Зhv3kؼm(_)cl >9L@*E-"jMFrk)dݿ1V±Y6y.GֻĄa,=\+"b/i>+DB[]I*ð]9@'߻^I~W%"`tB&+ h%5Vi$t/H0eYd|OQ`7ܥfpeP5jnleg]jgCx: 7o'O*ފ7)ݖIZgkHNz; ̻_x,F/kiIS,ȨA *"I![!VA;Y6uV<` (@;9mllFixP74m#NO'و#)fO'%miKI:ݨmhl"a!3T̘*J썣PeF:K$B\)ǨS[wKhԺoxWeF:y-*!FxmB"_):,L:cNLg(őrDE S mF,W_߿ݭ׿_j&Wo0Ѳ \pqagq?COܺ#h=4m1*mUwV́#9/AZ (A" ;EuC;⼊6 44PY+AIlG}Y᧏_b=E5 Mm6/fD^F*2,U82PX#?vpȺ 0,R= ,rX,њL 02%fĎ@\v͟kq'*ӄ : gV&i``X9U )cD AH2 ==A,^]7-fܾa`Ie#9NJ郑xT(XOɗKiAjGf#0l&\JETrD&N<):QX." Íզնkn9ٵ1M_o0hma@,k. œoa%'Xڬz#aokw_$֓zցDR۴6s&us6]G=ӃA7yfj6slbP {( Cp5 ,16j ʗ%͹Pʦxik@f3Θ$(3L 0DH;]>vIOa \s'߻=<oan8R9@;z3fqv3=uU;PqaCo!Stl88f&_?#6 oa#!Zq,@FdZr980Y<F&G֢S%]eQW کS$;|Z"D@+>G5tdx )C#.FJhTP=U ӵ%>N>l8_/{h*AU 7G !9N8@ƒbg"A`r=Pà.&C?9kFzbc1.Thh1E撛T/-" MyB!34mFfh9V%^wjIxG!޵q$B圜}C`i dcg_N f,Q Iq߷὇s'%gjJ7,Gj>Lv֟'$,󿊙՛?qYuM'JeDqaQN<Ig$*DFXupbyAEX&?Sgtr*E*ֽԈ"!߫HKkCXG(x:JuS%xFokv '\?ذ]$N!8AW7-mDtA Qefat?z2NRs##2cpa qg9S 3^{Aی:p5`YhK$u*kA Rm}r\Nvk/%m5}(>QL1[^"kJJc z%,"/Al('Py-wM)Zέ( fCnRAU|v61'Ӱ&p9fYwy.WD%^<<[3  B0YFv=~x\װpܹöW>wN&ÏjAu*Hn#* S]%߷l4UmK7N@IJQP 6*`!K(m[kߖspZfoݑ;KIۦ _QN2_[hSZ^{up8poYh{mXl!/|~=; l"ū_`ϟJ qoBc.#9VɫU _0Pg2<"rrnJBQ*k Gʏ~ꢦrEjEs:.+R= lU2LKjV .h;jєJKxD7oak| _a:I̔ƞeYC.j| BRXqyfjK-#VV V*As*$hI5?XrB r Tۉ;e?\t`ъROg&'CJtYPHR5N!nSvk e?1wOzDKz&k7])Hy(X+Ug?GѰ"Ftlh8,HkۊqF\!k}Ґܞaԥ u:(2`zca.J{(d]7^)`[`.̎doQAI &BV8D sGߪCXZ!hugfj.XGyv+XQ+S~ Bt؞&Yf!P'T$j<[W6QĘuY-j* cޙcP-2m^5x m0.}xeU4M}9Dh)1jSi՛?Ѱ+󌍄ԓk҉sF6%]=<%MoUln{#Y];w 'USQt ($7AE'_L8:q-gĬ>$qN03RCh̟?a4+KͧbϱF%oI٘quuk)cBEo4K]:/F/A,D#'^Bi@h-6אɬE x ȅ\`y?] xq0 c>hlt)ˮ`y t< ~Spfpܧg`= XA[̿x=ᾔ'".uE=M .++<DR2?lDH VgQWRKjH]gExкޒ 3$=+^'UŒ`\p^ 툞׫Y$Wg݈dgbL(5j8jTX{w"T3aW9ڈ7T8Ϯn;iѯ>j?2S\6YME׈ɻu1>Üq%yMR #T$Ct"]k%NEX!$hAڄzQ(#iBHHKVu7gb7Xc,u HQ2 ۚ$E; #LTXs]SmK1ADo6ѡ|E! -iiq/iE-5"{5WXb(#=1xv Iێu .Wv5I݆(^EO~4x)-=Q נmy"n,E_g's&hqtph(&DRvuSʹh#)br3fyL /}7lp4|ZԶUD,dBNzh%PlSCMJxAOS2J8MgpH RQSN?gDXzK30p޺7 7/*Ai~K?ōO>ѹ(D81Ϳ/~u:/~O,_ a|q2^?CE8UCt֝HHWJV&XE}h>(EL^0R9T/AQݵT# `kҵHlkZ9'Pc]N01"u'SFaёi'p縋N5!0"sNf(Jmϙ f$ISݙcI2ޱNG8J; +ԥ}'bD npI1[8w#ڮwJA=T;5Bh~yJх5D "ϥTZvT]I>;֍VUwRˎ5ukMav^feCf69 BUrP/s)e/Nj0'lϮ<Ѝ =M&xe y;u%@n#PaeB km:7}"S\ ,$ ftTEV!` ˟]::W9{?=Nx̩YAu6_ wwF=rp4p >•Bt丕%kv7g+׶ߊmp1689i`N1Q1o%mqd-IDRVV6_83*2ʱl_ʶE%uO:Gѽ 㰖p {N?$0RV8G͑\ˏVrG=|PLUXu1u`(RC e&ͳ/H*!{t3̉X_MhaPtj>- *:F=C \Nny x~&3u7Ss 7Zpj9icg4QZ#̥nc#lCFPy#Re L_}zp=|cWɀ!:1}!ln4g\_0R*=1#~xׄ{Ϲ6˒O%[LV*4#U8M(NsM(Ns/NW9S,$CBg;}J7\{'d*00"ϢUJ\هDV"YKRcFb u Rdx2]XwGq׳?ܦ_JX| &T\sݛ`J%QL:\̷~nʹm xL)Ƃk3< ̀D\~iD:>d!=f?:szAYe!Ԁ%;uʽLyEbK2))J _R%VEH%QXvGxl\  G1N Ԡ [,GF`A:& HnrHg%jY`H " C0cwдȀvBܓD`̤A|1SH  REi` @tNA9H$DJ@씦R0I#6]`׉h2O15 pxӛJm<+%&;ɔx49d探 pKBWJ'դϥƘ1[w 8 #Y~?O8qDsI sO!X& !jJQq'yR$ aqDTF#'\jnAԵdB _$:*ʆ 8lhi3QjR5Q$y5"L22x ʄJG hA|bQHh*5O,`Z5Tu5L@hFDB/5c5D k EFA!N]Q[[P8@Ԉ Q=!)i\*cF>'a,*$E ~H`"wDMkX(VFRIQptQ @U`%JaRG[roȢe"R'j߮.F-z$ *[ m+(`2|mm }`!.'E*FXՍ2]7M[_9[(z2A=hKLI{s[K@9BvH3Z/LK{ܶ,U-ɺ,U^;+zc^IK;@,9ҷƞfa鍂[vη^bD,L^p musf^5 :[PEeݱ:α59BBΘ l;T$`MFlMpՔUZ}6J*ZDj+*~\TG4==x~ #rDA̠'ӯyP xaI: &ȃ H淋&#/*/w~Q~/Y":{@a&/!b*ZM%O㹛oZFTm08RQ;n:Ř,yj+Ç$D_; ndu"Z拌a$8rm+#Xc˹) M"IDYїuQp:iQ$bʯQŘQέ\4>5[sb[i 53FZ"ە㩒*02In(51Dq}՘5X[l)̝kǤ}8RpAS- |뷶3^xFo wWSNF i)#>D xn,!H Ҍhv MPM'GoqfIIw0?S ".E\!kq> 6kƏ"_wa@߫_e׊v߱?Hh% xz݁?,HE_~k}Tąw|{B0mYD_oJջOkGII|rh.O| 5o&]e `}R%tat %ŹsVh"`U8Z9oT،0iF&*!jL!D6ydKUQȋSM/M/G#u ȞxRϻF6υ./:?Aϫ+J|)y2H9Y5gJvYJZd]YW8@͚"ݫR\nUؚvޣOUypyKN' hˢ`$m}+V[ܷBKe9'g-MɚzkgQ> &5 SF\aѴr8c1+ zdo\\Cl,,ȗ"Zbռe@lV zYxIGێ.+!T.l@P[GŻ* ԯ\Y=v3,TxQa/+sDh\LjصPX9/S1Ƣ/is9%}znК˖Ӫ|[\H+$V9{fK*Oʿ{i9y봸mu١΂x0} o1?G?5Lk!i=jɃ8F*&cY"&"K&KS&C?ȷDKOGb  D J$%Q 'E1i"fF%$Ox)LX-ؕo6a*@PX XD%f68B13 T,c$$117Na$A}p`UzV=YGJǪ']]mpB9` @tw[جj4TTf'*2шX1F)3$aB\EQm#DATԒJ$6҈qY%8XNރ%4T- p!`:jV,FXY4f+!$jKunӇcH9vv:ߡZ`0Ɯ`|/}!h?.?~ݢAzg#Hֿ] `^~e",?x?|ʓ$$%"$ķbi)L-aBHW$>TZnѿn|}v׹>gb;o˖GpJ#ufgIdHQYl8#[(-eQ\\-/?p垍`fxb33^ܛ{2W|4yHإNm37ue|X7@ *us|OUn;2H S)[Hw;zl{L1&8`gғ ӃXM$E4H/p>hP N;hy'Rʣh[r" S 1zj>Zv.vb盘 *ᆵv~jrg/6bg撲N0& 5ÉB?\ʪ~[MStZ7C2:K$>d5 #c|[Bd+iXsI>kD?lIrž=ozJf}3h6{z?s X0tdsalȕw 䳜O_ܠpcÌt= ã"TӾZ]"dŒ;mԼ,F馼, 21~pC͕/0\S0PKx.LҪͭ4f[wLҽ՝l;,U;F.>e0<⼼q*՜͗hz3z27WWy|wB5͗(g77 | 7}Fy2ػg߈ C X&K,M+JoՓ>~~={{>K%QR`}88u K!ƒD(hzDRp&@)ErJ }|r@!Id25:QR,ℂ;*F8V:?YH!Zgk>͡ߩ@Y!ԭOFY%CS,"N[ ADr\Lqd. >Fi:S\. 3ڴ N |{ś/.aI04{s)bj,>l0I1( h B>I`A$&$b{s|0km8"=Rw}>'"xqB_ghKBRc' &)s'1%:]U}ljE-5gea8^@ yǃ:QăU$NV-r#wR cUDzkOv22^8}ZG8Fa'XWVxOM [ &Js05y~P,[YMXB n e(eF:)& Q/X[Fڠ6N3Cy M_*x6+0):bboQ FC`ބ=E$Hg3>X/XXlAS;UzF N8n@rsRTJU/%J8nAq]rg%,|Aw KI3 V`0.j ;ESզȊ"/I~NzbYNCzHsr-*-vccY=wt$2%`zZR$"@aHpA6hJ oVHI\8GD".GJ/u@gbAV(MQ 6":aC7N l` 耺t,rΌ4 Ddvkg,@a %.I5'<7;UR_ %;S&j0UAҟYOPʨTBI[4F14 A&hV:J GCxK$˨ BXmtsHHPcҌ(KV1a <ѣ>aޛ$/IM@%@8S)ߚFUaM~u!9t'9UҚT%i b t"=5Mjmt!=CDU5K22Z]7x撛^%/6]4&[0(E.yP5^ҚQgPBJ/,K[fXvA鶼m#sAPrt˾B ZƒȧvW+AsGƄk4bT)ܾei1q;Wd9cR/=Q+|SJX!%-ߧ\ى˵P7u 㠧 jpנf[{8͋~ڂtNn qta咔^8Inք8ZCh2ٝ!|0_hBĻdI>2.Gn>'@F<7d JQL7&࿾Ůn+h.Zv*L(Pnwecsmc#0R-*m"gΫ7\7bZ`GUXiQ5gݮi|ݯJjvñG&ΠF%tRjJ_,reGgrS;Oޜn_J Yy,`0`OC5/ws[+lC\B=pn2xL4/c?g?.5G}+8=8m` ԉyzWB=̤Zhfr(o5ryku rrAor7=NPwQW䥾$T IjkqD=5WRVI4Q'Bzujm~7Azp] 8bWoW7)*gS>kmLGl }WZ w\P+/^#VlsuUdtw\g7 rq6!--TtJ 452mEӂzJ̫u?{8FioC {6zlۢa`Zy!eGgf!PdCŮz68ѭq:+=Lj n` :ϙ4+8\te٫Έb'elD~dޭ?7UI&w:v3M7! >s xkNoOToX@ʩ0k)4BKcIDDOXJ[HKԂ痨3(,bT=U\~4^x*}j,@jEI* [1}Vͣ|jr=H n.X\0BzN-QΪ(td$EIhՔ֌]2*ӮK UHO rxE.Ac MK{+[Ȥjp|X Cw2w!,}h&#YP(@t KV%SR,axR ǘsK$\DQPÚQݺ]c߷r lA:M<Gwϋzpƥ1\{i^\}ӫZM}%|/Z9C2B0" ̡RJq'NB_%Gq=W2Zxk 3闳t~-\cBf{^+JuJj?,e F E ݋`ݾ#̵!bi~[ugDm.:ZhJpoq!䐃+`:yF9d+P59O^ه]gS ]A놡'{/?u0﷨5Fyot}d o$ku@XazTfj1~Ts_Up"+iIhX%M[+LGՁu %B/-.ޏM|I$|},y+Zwe?.n1D3[lee~ۆp_Th/MQhN"=B\Mz >((CT!L;5x{ rՆ mVmD1"A> y|6ȿ""3'rܨ]aWD|g7)jϨ`Wgo(%Uy;7sv쳻Ӑ1|usiՄ4OB6%E4wXbWN1`FiVjUXnM+M/ₚ;Ĺ/nьDV:|7Ui(6ԣU\ҥ(Pdhɬ:'уՄ3)CDVm+"UQS/"uHEe\'4 #-LJORYJK ܘPf/%daA^{|pz}im+:}fڣm }M2 'um{΁΁r |m[MMLHʸwNeN+I@g%rb2mյLasr)ӃdtRʰGQfȞ~ T9B7-PvQOؐGM5P)`3p| juZ>SHQ a}b:6KR8^G>@! ˡdVuPVn[Tɰjm#cFi$sznxr;GaP4\nYP{΁5(Tv|a](A4E-Xb:ӆ(9! אb 6ZtMjZ9'5]GsTBwWfp6 ުE tJ$Peu@-TGy?5o-/n hԟװ]eu,,Z8a‰S HS#(iZД@*(LD1(KhKRB WT'.B@8V~X,-#-6T{NqN$ #^ Q6 VM`#gaq}+*Bz(\毑ص9akX@a)N(U8Qk ~ 3Pch_u@-(sFDjn ɫp#9ap P`LmE~uqaJϺ},cYƮ]euT*:]b'eR1z Sym,qm^tAH-`rmۉZeݶ )TXzu5摭O(gm@0Êu/nWB[qvJtԻYĤ 2UY3BmP<=@*%{S(V{H6@TDV~kh;&ArCаt-S\X!FN 3b 5AݱY$ dA.Q!MQXFL[MtTL.ciX!anնwܫCN%-V}<@C_jNjGgR!`*8 fBbJ)\2<"u}M;~T 3_C\=N\QU_;ku?uݏchZ=(.y3Jdm;FSo>.OEoN _*^:C|Х[㎛r+o r&--q/!D7DhE_NΔʸϐj["&8;˙ =CWS«&BDRÐeSekp3pÄVJQG;!"矞2K V:D0_d'㐝B5-#M&tjC" ɺۢ1˅?d'+ISx9RcѤ+ y6L?*jz.ٓ^c}FhB.`O2Sg0FqC@MaWxok_`Iri;vc y"v,?j|M(M$6r}ȗP.@-^Ç"rGiҿ-6Re%ag49.fUHҏ C]mk#hW5XѦ=6T"~;u F#ט{Cfh"eTBv{\eۭ8 Qad08R3(q-6MsBp ڮZAѶO׫֌ jJq;LM0@kL:nv&rSif1SM0B|hi7rvjYiBu2K 8x<| &?U敖?Rg?-,lgϓb }6Y&$c7=~eeRo/R߱}/fnQ6^,< 5,ð:&do/R;g?i)_=ev-|&JY;MP6閊AI}GvF}x*qo-K^tkc|K4aMiMIVT!zUbƈ (a9%ފ< 3F rmpb9n[Y{0fZ{[ &ywOm(^yDžkl'8ƉS)_$c Ji+.JIE mVa>{}!~5;KEk!^%K-҂5/{Qc¹xQ_,4a8|b0g " SZyokx29ջWLh +kܺ;ӏ[P*F:KF9i`QxrN1LBSvGkZs{z\]4z! _sy5L WRJF`ĎsX<uz#Co{W,Fƍ֐'w0 +RX hg6RUSc>EwgBi4?},Qhs%F=/2frM6pxs&n bF[&aq*|ׂ)k gV鑻7QtC|rw`1Șy|gY=F?frLӇg?#)~gb?+E~gAagiGǔ[Bѯfi?~? u Vl-XQ[\-|k~Qi97?lb85B-|MЕpbYᴑe<T)NB]&x,W*s8g˧gş,plL`yn1(]aǻa&b!ao)W9e9<[˧I w ƅ[9mxN38qp SVT Sc*@`u-KCzzI-ah+VguߠEps[B>GfĬ(1߬+=!JuNOv"ΐ>U5 ڢ ng1+3`AC~t'T %;h`Y4w噟}6V|8}OFB'ggNJ[:A]'gWkd# qvtWߎK.1bJTRm_cU w# #ZA$ ]5tۨsrYG}$@.$vXY۰(sC]*QH_(GsCQt%DbiaR$.$BU2+Sj5,/^@F h;ެ(ԎdVU=;:bPu}-rKJ)cZ|;NYY(ؾ][ =wVҀ)Mqy+ XRFc]36y0sO>mT~MΩ(Jä^ Y.UEw>0^gE;j'A܍Mhۥgױj*ޅyzNU?gY2 7w7&i-e'dluM84qo=&&eboX[d1Yu'7gIaЖ]maDž-n8,gFZ)JF8$\)˨SH(VXA4 s?8쏮@bRorf[>ܤKR"1s0m0 TU{9ő-<\8XzPRhzu(rn(j Faş׈j)Acr@)b"0) pNpccC!@5eGkKc9i>W\uRÃt,g2cX2P{8xUǼߓ)Gɡ٠t< 'xa%3<3HцIyR X˂v~@wAGI -2ٝ%Sf˝ &'\5՚.x3/xl_]NCQ/Uyp7upԴ!:W3B2qb8i_/86"-h҉s/TTz*/Z6-ޖ\F{S'RP5Z)cU-#tSءYg "  @ཆ UҎ`:X_'?۩_)|Xvv~>z@% 94W^vaZKJYm*v8]nq=}B◛G 2oʘ},zՊW&w*գmyRY +:݂Za54'' `w&MAfDlgJ""6Àxq˾d͗ϱ^l1bK:,-Kc%Ln_,Ū yJcN h3[5$(x)u+˓߲qu*m M+"'1Fp*)S^w(8'LBy܄>Ɉj|5ZA2\M3sTϨb9$GpJ0%V%:5&|a*Y173c'dg!OJ$S*y4Q6sxa<%&jO2U|wBqW[Og§XJ6=]*^prsTB8fTyPwr2V7 ^S?[aC^~mk`B;T@Fz[H/o܎ɟ;lgὡC;Fs 8MKܳ< ![*9rn ٯaA91#S!X<`FoF@\Ydsd3 p{?徫]uL?7_Y ެŭjV62|w{$Fml[nlX)T4l?Ϙ 6yvS| aiEYjbe;ד;Pi`B!rb#/׼wFaJ)F ժ6n 0mUf}Zȴ5wkh7 5[t5_0z@;Ag0P]92~H\[?حkeTc}6A۶w3@hwѳmW}Xvgp? }*BF%>[D"]5|A*.)ZpfS]cOC1PMR 8{Q)Bw *0*R8VRœRh%p 1LsU#BIS.F}W vx >%eɣ"d y#J:ENb$R:y 9A[f 7 X&AKXK٪Wŧ6ćcƎ/F鬐*D&jC !9eaH#\7k5cyխkkjQHT6ء^Ij_NA 04ә+oS7qIKpV_ϻeomno.fb1Z?x=OM OqQt(߹17sK遼\GpB`@rD Qiѝzk X,ңq5c'%DGTK=G$8#*rZI(Q2ιDx e%(jfB=ec cJу!v3 GwܞK[|#`*mΛ?~1q6 &/C*30:fV.u)r V=vkߤYOnD;#zloT_:o Y?̑.T,wf'Wf蜯17[A]{F! F*TĤJˑQV̨ʕs5XG&V,lK`֢ j,I8*=CYy*orPK1>W&}O9*F=ګB~QBu>f\~洞]h8D%Ta4?W4oE\'W j;"PgRI5Olݻ*W6v Gm\>fߚtacekYq5iWh mpxdN GYV& }B#lgpr8%>M}!|.GWm1`BWRf .$0"]30)Jļ9߇10 B3Na2g3:\\(0ތЌNA>'g)jMg:+u5ObETpF+1jew`,rϖ:Vk4٠bR/05oɁ#ы}eDkvJ>?Zwt1C@09Zu5m) ħnks n.FO e[@V2W9_fa^ S5 sZPps*iR]A a}zʟM"i޴EQ4AO@:gYi; *!"`P> }ϴIef$ۼ1M$1f GE?[7ݚ<ެ6lHJyx8fq#<Ԙ(4aRulxR\(Z$QV5P˂r (rز4T|j%i~ȘLaoG L]pAtZ'ʼG,\k6ax9&D3)K6G6ur3{;ϟ ޖlwO^\&\Rb)Z d'@' sn\+")mzGWy~#ފj#J8p|05ۅtm[(Ķ* C(.dXm8'> jJb'\(Z(2hAƑx3+ywA0^by.IDԾ׻2މ#*i܄֮>U.|P?/z}S+3Is+}]Ԩ@'g)Sq{K ߓ QkDٹQmNѭlJ0Uk(}:5eu=yͪd\VBgʔO'&H3~0;NiBޏ`0kNE J_0HaxКvtņ>V&9==7Lñ^2<AŦiMo#*֡b#&|$)ѲUqrBRܞ֫31*p.{![\0UB(Y]\ r:ݶdnot_G+'g*#d V&~Ϥ,X )v8ToV\냽8&a}SSŭ.LBy8_QAS>ƨyfyB¼NiU{ļRv[Y|6fi݆>hcX445h\OF^)}]4wqBJZR; I}8q'Z 'mjWJb$co.4U=AOI|;BvMl&z~WWv:ig㛾 6'!&^SKwJ{NiztatQꉱUZyJ Hp+qJ=jf\dT[)CEj4ɊO%wvtP#0RNP2Iggr*~O;}_g`O3^$iDR&+%Ojt2 )Ή(ɝ}쀬}~/R~/aUk1R``Q#Ք Dn)^s=S̶u;e-vlV)JuC][=۫24o.[_|׻jr8E,P6pL* ont> 覿&z$W0U )J͏VK@r,#mBL[(OdɌv G]v*0vYc(s<@I1Fq@}X ,9 @`BrXĠzP2(AEa,L"3 BD>4 & 7R"?XC,ւRp[Ӭ&tA#AAjG5FXl0PpM:Z$BncMg!$zJ$p$%LJHK3K8CR/d E, QJkG>,Ѡaa k`0Hk&e Dy 52Y rxzKW_+>2@ѧy?%{wM0'z{[yaVZMjK)#W ZTs}Ώ?vb4~Z..O3|Α~:uk-ݹ vV r]n W%03#JDft2Rv"RWv6ڒ3@(/|1ҍV뵂HZjt?ฮQ!)hAZ!8M^:UF/O?{Wm/-*/_ (p{4M 3$ǫ-MI)idqhl9<<$;z?u9e9XSrGcBj op*?䬀Zf:ԍxw"@8bå&XQFs{f"2&G|9V16ZuL MZ['_J !J*: $z$O8YRZk^T[!"{gtZqVZ LvU Rk)lJFM%-UUl?>MWͭ{b 8ړ-2pn<>{hLARqH^DDQ`%y㫒TV0dhIeSE)pC 7@M!J!Xr)ha ]5lї0i |cvX{6 uѝ=ȊUe d0&T[d\@ @kEL'X9;*S)(eR\1wEg),QҐ:G-ک;YVXj(0M䮆L+,+a0 ڔE5ʪ+T/N)^pG`&b4Q(=/X J[7U9"qF J9V)GK4+@\ RiVĠ[yKdށǂ `w&h#L%Q]UHJ۹jN61~--o댎L9Ls, /K9 vJz!~bHS\RR {Y5u%?*maaS`L$4 ;ˆL E!?Hg|>Avy=1;jtJTj6_MY/ EÚ^6^ϟ*bU VV%[#T2E Ʈš+Xmt0;%aݨv$ܻws;JJu?4; 6xыMtW"⊰ze{ A;+zkڔ_X_(|<(n, )Uj_,,rqp!B-WpYsi et a/پ \]|s=m ؼSSqsOFP4' S<6?f1cČjeY@W1j1phB|8#8QBrQۆƒBdz߶Ah%D5ݫzٶo}4(WV!m#x<aGBu;_KX}6V2nА7n^1s&FAnшpn ?ƥR J܆m_6:}=)8XR)5!Б5hDdNn;m[܆%Tmɭ|v 0,{bKa㦐SMu jmzrp>m->}}z?9u48 QBWH萐t> / CXzvazx S"#R)NaüeFm) zx2Ĵf'<6btv2 1ڥ}F0/b (cf0Q.TFF˄[QΉ~~mR1஻uo!ӟd3up(kp؀t /J9+8%'kh_Kpx 8 #;Z#Dw=ڄ_Lg/)y_2vt1^q䜴 shیc,\Cۥ1kG18: ɫ0N~"ztQZ3 ,(h#+d+)"xbQ&:7gU2yvyPWz y-"tI޼_+=1^xR^'n*&}YTv%_/Eyq+$WCA_7 r@͈֛(R6h\Jee}?_L&K+[:b;FKm} NJ96XO$hR ^NuCQ@a d;XEztS&So FT[ŒӯƆ_L}cJ`n6 Eb1vnI[6]LQtj|[b'߸@7}H̯O*[}" +G D^AOA0;mHsC>Ҽs̏C `nA~cһ{1雧,FV*{e~ˆG s,nKL˙m&;oS%Ɇs[v~3<:}2dv`DȦݩ ||GoÓR^Vǽַ'r3@@E"hD OqI{ʒYwbve%€f:o B\a_1an/+Y $-LBcJ,",`JURi,.%)K1/tS  eFv24<c`SH'lU*:)!-VVr;.EP%hwTI,l޶qi5{zoHRgA*j%%..3E%tu =wRS C1#1^O+g LlIdLXY2MsH et4exk6 CQbY5A!f[qB< 49ݯ@Rə%/Zwy!hu`jTy?ob9_ܙ5f Q_H)G0ž׿4).E7u,&Gg߾ys U/ugM]%col}I%BbV\oJ3QaUZ?wK_nӟ{91K.k9ݺM[ZW*?` Y.׳u̢7Pj LA}E4 ]h&"XnŴ^ob9"qxݻ2P((%Ea䎌1'([7P]^'t1xh( yP ( oFyj5RI `X3JŚ@Q*fAg8P>cƞ[C0DȮQ ^=AG7 nGb4; \h]ϗU٨Am?S o9LGtz) ̞4_`VNQ٨R=w}h,>#5BHϕ%JW8gËwRzVwǏ(,'koD'Vkp1_vIicg]n:J&ດH!'%-$ ;Pmic kriB|n'FGVvO{fDs"Y ?lGRC1/~@VBVzhI ? LqE3I .6y:;b*6ΨTQM$UWգ62OEŪqko[&ޝOd=Tu0ĥ=1 @ z0[]ǯXWzP_)aw8(O:b]UW?Տwee^54,"s&υg  +|As;Qu|q~u=w`L__X,^LWO;ʬ~"I&ֿ?Z*8QӳCs["G@^QNؾٙsc8Iv3g?a` FgZdI^Y`8$[%_-v#xp=[j/0&80ptv5KX{f%mFz~2!y}o8WOg@)0IBo1Wm %&eyo*WG̖F/n c ΩưΛ D\5<A=k߁>rb1ήyp1.>{Q:S/:gF7>!~ym E=/UƊP4. >5yfҞ>-W#_^;]J┑R1n@VTS2E Ʈš+D4}_;T;! } :/cT~CIPP$rhZY4\UJUW3k{RCB1F/cjqM]I,.J[.#Ȣ4(sbCEXDY,+ b\`+0iӤ>- upDEUqzcPØtٕNz'`}R"ΪzÄ )dV&QXQTp)4-y 0YUsX1le ^2'dK ? &7]%F,Nq70+fRIl`(c!-P VеZ0c%9+HkAߥo"0)Bt32-#CC2GH1uvO&PGk^ߦ79[;I dVay3{x("+s t(1ldd)`,$aF3gj &tN*Pn19T1E@c2]jZM?\lz'/UGP吖Z Kf;iK0T=D(Ξ΀R)1'ߑ;F-$ey!gdTt^z}y?{_3&Uj>zẛF~ d_1;LPO;b+m: wm\suhhghN`;nsnN7XnhkNᾭ[@C[ U4FWxM{@A~u.dT=PBCr).VIg 9U(6̑F"jϷ!߸Xo5s !ڛXtcA/6`3C=~|~g |M^zCknrf,DVi' $r IRa4F;B †qߕВ[P2BJcpa ZY0q%Kjx+^ƑUqj K+V%R\ \ֲQX+M9gam4UjWW81x q ˊndcz9*:)`lM4A}RNdЄ* F{p`lלpX/:E5iR$N!!$il( NSs'$ټ&Oԡ=4L;g-?,-nƣ0m3nY:,,xu6 Dvџ {uQ289kx|(`%mx1L$P&ky=U ġ#'wx˙ 9ѳ6v,ѳ=i| ;Pkm8F;0F$jKJ@Ylb'! Q) 0TQrQqEBcZF+.,q*H8$Clsd)tp-kj_~|< 7v}H퐊4$atYm狹d&(̍=ҢjĚ3q1X@5:K.T<нμ4 7_>yel[aQ.xoB Y~,W1Llqn%^n~G޽}i +ooJ e¿ :f($ ӛ~/aILoB-J={Zp5CȘW6L~o0"TpNh cMHLXabE6b7rM3A?޿/?A*ӷ%ـ-"!CRF2:*HHh aGbU<??;T"ll Q AyaMR չ P.<=ՠ"T\Tu$U}+B/HDQ*.PTJbOQӨvSK[-قyp e VawDPϤiZHEx|L'ݍ"R%r?nIjq>_rf9o@)67{AT+#Z)-뤪mZVը-v'yK$9^Y@3$3y/Mz^f/{r6mt~9]|~^̭.lj_Mr!Ec7 3w(& L<}3A;;>%!YPmV0ΰXBzls@q*\IO)9VjW.Q2š}XtAbPFtbQE1yn %[E4J=$w-2*AO\w-dvCB^Fɔb~k7A;b1(#:諸:tƄ2O)Z Rp)huT wM*߿,*APo1Ta Su:Ey򑎟? ^J/lX/!.#Ljj"K׻H/ 4tRh6_,?vp) vpK-W;?cEe~brjpnB^@C!jlQ/0z+y@?֫i Ο.Wcl;Y_-ESJǤ@K}Dv}-[XSj@J^3N0BUWeu$nX_& Ǫv)@$DEtpI)Es9ڿPT.flR,2d2 @]D"F9&V"A sI\'Z+bp3c!$\-)ݟV u] 7cC7=g~̪׆݌Gf ynޕ`Zrt(Q.JTv.0IiXMH9ƘTHLI_BG{qWA\K@ i!1@`' KLRcc9)Rj#ҚԿ_J+b!hZPgY}m4{Nezcu7t1yrsFl%2z֩!r9sr\w L7 ;=^܇Ldǽ}="Kazo2Yyh@Pi:0m <_'ylzOMo1Ȏft H5`& Heʽk#͗*skNX΢.BrJ0z]'`$vZ@P] _ (5*YjifW% )& R7oO8JMMLj NFMXZ $ˋFh,E9sHs$R-.~JCRo (B۪ ZTl(mt q^B&Û3Z b@pP0fı5{T;0F$jKJ@kL鼑L~x \KLy",Lwb Pק9'!Y;0_>-ܾ>};+h(w)/h |>F#Ac~)fY 2lPOŗñ."Mm jH;DdY쁣 @~~'goL71 _zyήĦ5%jd틇hhV?~KX\A8T>Ȳ[g`1عA^Skn8˂G[gd ,ʒHXM=̻3g`ms XtMZJU(Xؾ(&#S1i֢lr 3<o N1Ep*Ss\)k.\H%&Z$VUL ^Du@۲$X|'=Jw+W~m̸ʣc:Fݻ]0ƲΒecc {X֭WDtNM a37FRzrP3qyIi׵W:Z*dr] 9oN~*a $Խ&#~|W!g*#P޵57n#뿢˞.qR$T N4kKd*~lSڠIJdj<h~ht7݊iaQH('ة Js^# ]S=Q*ͯEZ/]1~FF"٧ˉ fyCܙ׿w6666rw3΂4BQㅳ &[|V ǬeG6+Q?#.b?J@tvѻE9NZ'(XM|ɟ/\Vfs~#EhINGBv!7Ɉ"")\{8/+, 9N L\(%TynJntj%AyoDzulT/w܇wiVrV~g띗'v%S| RSù^˒%c*" %b*$?wJ";9bވdCSÑ']Ȏ}П>Z2 dL>$휃ia՞:5$XqvC^͈2;`T?h{u 2.urGD& {YNg96sʊ2ד^h h% dPx /G?6.2/!&.h!qHE*wd:颖'-?i Dр[PlBA:aG}M _482A4x]-o(Zǫ[cR{_5[e'adJ JWh Eq1#m@FD*OgG6`cj+"+ж}P!W|"NEOÇSeD+%4uxS@I/{ϣk QTOgWv{?3ͅ^d5E0+)D隁KrJ#}'оH }ghci΀ˡv8\]Pǂ sE7'F^I |yq7B29ӵ:i.KEǤg3"I$}|-0Pzkq0Kwa.iGEyi7uhPyK y]#ՊSk]R'ЀhNv' =LjQutrXt}7![*vy@ް^üwI@h 3 9b_&w> s\ YVw RfyrUW&Zd"}*Wݘ#Dya ^MdGB ;@ >(yq#M'z y,H?}.O}v0}H3k PRi@(}MF ijN14U18/@$O ϫ 0D+w"PҌ^s#(qcHI*_*&ZRݕ3Pۘ]<٧yV{Zi-߄^N rA']{ƭ[ 227 (2{iBiF4W{gY);W'+ N g;hOBۯ4&%+ {liO7(±9i?3ԾmB4C"?6 t>"pFF%_Iz:pk<y;~r_cq9\u&(*|n4p8QU&VQ:(7'Q(x#V?qiCt(xsH_݄'W&ث'0ȪN? 7ǯNpú͗\y{{y N3"SΥͿ^nt߹^tn-F /I9PqnNncޡmP<I<_1LD_ddNHJf 5`e-c*f]tdS. 5^i}eE;,!r)TA9TKhUp"pʝ` 5 F*}V@'su_#\v:%0wD"Nef4?Srp|4knI>s͓r7/[cxַ壞iҒ tmNH)O$eP1h7%띲LNUUZ04#`=4f# ǘEXp`VXhF  0\]nM'uMߏvyfEÅ58^lb?4/4(͵q[@!hQEн9zy m[IC"Dr(H[¢ nkN# Y[IA1e+upi:h#U 8*J-iL8R\[! ѴAFߴhvxJ-Ł4Hs%pP  =IvyFl8?O%43giI:4k>3 lh*}!^HL#2!"0*8nT6£ *]n Bm^nAUfhr- r[<*)F6ǓJR|0%OJ1&RHBѠ1L7%[(T膴7h%{/y[kB*W *9!gAw :* :W?.. eU(4@AJ˲ZJmP$>9<_!ʭmiw׈!h\'bȸ+E 7%0vm8!\] TwOP+Ah!X3Z-3Vݭdaqk=6>mаkJE0wWâHhTkFo§ o7|v.$<>BbDkLB`v CUd72Z[ϧPMuTn;J,}xT)Iʆ9˄QY_Yz`r)'g}tfS醽?Zəg΁*SqZdvSzV.1;W.dRj:yCA[[ RDM>ugڭ}-Qu!!\D[5~31>J `{PLLSʥQbwJP&$++]K atٯyd]le*RSr. /T*GWj}w\l6y? b-T`}k{M'OIo>WyogCU:Z=B?~oBmUdS꘧o/yc"uN_k9،!w|{doHP ofE f9USD_5rs4Pϳ=G04ۿOo&Xo ]VUd{P\& :OtYV{9iމ9' v Q ο;H& ßf8/EOywcuFy|O_` 'ޡӃy{qUOn˫eWaQW˾W~ycT&%8 \ϑ]TdߚQF;UIdi-åX lJ?4/4J`@3yKd/E̒ J(VBĄڲJ 2!6$99f%R3*^a*\Ls 8Jd5HjBYNwt `B"̏ډ/m.meu`!ot0\*Ԕ`X8k.aj{|LqKF4e2a7^{i b\o)ʪG [/<`8wbDzq.EHoW |*{8U 1žrDB6zb!BAI9Pa*-/_m1@8E7Vn`I{~G`F;[, WN(H8%(ppOCP"+ J,E\Xf."BUߒ2eͺ$y) RIb\߅lu;|I/Mc3Ɓet/gvȮbIFׁyc`3{DCS1K'Ah֎t{qQ2K2s\p\5.Ww_޵_q9Q(4޷  `D{ʲ"`wnqx{gb'-x?xx TA&k7/f]3 7j8`)3O]$ㄉTyەŝOk퓃ly;$pVAtEpS~wQ8{{)lg;{B ZSV pV׮yXnBB~/i{ɝ`/pE>~f6/ z:^B!yx@]-RMHյdE})w `nwEy_}24vRі@wAK5+ 8/P0Mxt(wb6l݌7jwI[/ } 3Y71?p)Y(O8h'wxe-=8V LI-)[5vp q] (u֟qt;>}bz۷y8@{s0@Fpt 40}~͐%Cb;+Y!A17ep:Bm ^@. BElo! ް7>̿4# (@LfV;bxk#\"F9NҊ@{VOQO1^wmu('jZsѱhփ+]c Uk_Ӊa0e'e-c֖Vmk? 2FAa:fcF|ɷږpԎ0R m8WB(qjE(Ie&%JU Fe'Tc,I] q<ϕrޭO112|$=qk|78>a`F@9s0+Ӿ*x_!J4iE0PTiLs 8Jd5HjB'N0-dң5y(xߓ?<`λOW ii-h_]t6_?=։~" Boѳ]dg(xwd* ^}zrOTb뷐!8<2_On 1!"Z$_%H-)'Vx庉I~Q/ŒWVϋual?o_sWLO֘Y$w ~}/|gw~}Al{.#S3En$#iQU9t!R1-7s̎"6UwSb87Uc̓&1*Typ:d_FJ~ Y6Vʨ<9p#=}$x&g|=qxH2Ɇ( Y7`jݍ5ˀC*?&y'_~O&nzwᠼX xqu^Ew>\dLZ"F6ץm&0̅SvպH&a8B-5mT&kNvҬ𱾰[y8 4fdӷV&@ǺCegooEkUߓ1tK/Y>V5*ބW~Zo׳tiT7Ӵ'OZcQc YRX?Ee}q& y&dS0Nx [*1Gm@nZb3B-=3rua!DlJnRfM{0$0ԬQWN8gt?k=EX$! ,J]0]QREŜ_NsIB^4X*b_TvgfG(Yp 6#eN;GQ|8u]9#lH-t#l̸,C%Gj/rs_|F'_;6 7Pd/ԽC_؟ߞ[ {\,[F P/?/G}&7vBy=3=}Biv>Z[xfϮN)\f!xk3K kɁ 0-xU" [d\zGn)* "TV$Qak,F 4JhA'AX=EV;M$TFUfckTF>YZ`3E}VQDM}|B3)90"Y$E#~Uy(#;.Y"")P 7I(6V~? +eӘ:@  )bP$1]kQ;\W[U-B0&'AcKr2T`6X91bKOx3Nz\͒] ]~܍Vny#{ȼ[iٞ|zW[Rq"-r'(NE%pJ(R VL~.0a#qo%yT 1rQza_J E}*1Ak}t?g}e|"*sX+76%"mJ\ZR!^WGxp/;NţX1*28~C6K7nh/>c[tA; ȄXZmaL$n3:-ַ~nrKN ~ (h^&Rm5fbz{y:2 sfOG'%I!&II6#ˣ|xlqZa >ϔ /"<- -.]|kڭ#z!9-I(@^vB*pϑgh *,NHU6*Z|Z~VыppI0>|cX6Tםۋzzus]^- QnW˾S2>ycc ܃5'gIe@@Riр N5WK87Y|#8`I5cTNʰDW`Z c5%-ے:齖X-EBE"߶"bO :c#."WL0X^K8cc1塠@`%7XeK 5[xL 2{}s#r/~D}~9_%pA/iD#qnh9cWXc0^DliiVr$VAp&3,k=0V>;Gn*:NϠ@U=Ī26T0, +gN—W5_"W|.E.?wA%;[^^R@7|d?e/yow].Wԟ]^|/Io_%;Pzؽ☪@ջ* "hooހ73גfÏ5th~cIЦ`[hlE5uG5 3֝ PC@( *02eh:-CP F `FH!u[%Eɠ:P̨(ζupi_WЭMnjMVUVbOfܠSv.M'J[:9b$LƮtKsԊ>Cn,Qgs-OYdvKStq)F(lFqR)c 8,(j1(XNç:U6^6B޸VGaǞzV5OL?5m^zعyik{v.=vdZVSϱudGjNXndި(}mUh Xp;\?M2pkCE=Z\mЪeqmԃ.R]D+;}m6 G?".cmk{K+Ҍ+e_(6B͹E\p.B(Ğh$ׂYr]1ጼy - Ló9#v?O;DE8ɭQ|6}W_!-scNz&.Dae} % 'k7uT}_݊m~wROУ rdU:,&+)PY"JղwY:a*t_.a1gȕԣA~Ms0/.2[5(ж;gh5c3,Bvo04KlňۆLb O|)ns_w)^=i Xك ,W歔]|(oOܿ+-̴uokGЖm> 蛫'~1O@WJR4,\*)e]砡밮v^k;"L?q@Nj\qwFˎoo੣Ǎ 0^ $;gAdk\;&"2@16o%'=AT P^zUpMˀ PgC|!fBv UP'$S/XDWj4XqmԀ:Ha015Y"FxιR{JA+DVPͲZ#Rع&d|4Hm54Nmw3ԍ6Ȁ ҧT6$& )Y6Nя龤GioC]]d}kxT[!f!]=\`biJ,]qge4(sAb-5J C l%MGvO {ڽYzА غ^]JCDZNw %GieCFwMLU@}w;2I6h5dj+KrB{ԅrNzWJSGI4c)ӦJ/, O=xtU5mR6Pk+2-PCM8~7 ٰa `+dSvGhvfZqcmPKPX!a*njIҫEKܳ oB^-P{>%mRc NQۜXS~<}}Zd )Mⱦ ,Xl;.<(  rPI`[%w[9(t:Q' If>tC=2sS}6_턀wK,3"-̄gǻ 3VA>w;^T✱[}B6q)?nR+*1(2gtnG;[#>wU(nX7*jx|8>VA>w;a'zlVm y&ߦ#t PY3P3翟3P0P`=h/jfy?P[6o>ɇ6V+ѓa5o>}Un`*Zi``T =B؂XKo׺ ,;zOW9!m/жŐ6~;ֱ H6Ղk~;[}|/׫OW_;k6%+E-7WlY~qyE?p{ skoO_,0yCv:? DžHCׇP?2O S.C.C!oxsus>C\8zs=Xz9<%%IQqMwcoFcxΚL]8~x>Vc˓Wן> UF\@˅Qk9jw?}i%w8}v f/_Vu˗ܡ`qv6s\$dC[v/g=ުmm1-db̥QiJk4`vM2ctc*3}%DF \3΃2E 8} mj6ɖZsH,i-dUe%88c 3&UL*za8O(ɴB/{wS:iWTWnT~kB eQIhK±(C):B-sKmVj&M1$2Xը `\DZCht^ bP*-&rV&X4'\ AB a일*?(C[-JZ]ӨD-MY"=~B\gɈbo.mwq>CN+ {;ۧ7,[Hv)NvRkƋn\Il*. gIhQށ>K_O'}|~=0˹x\rɍJn4=T@4[c-9߹M@FEbw \GG Rs*NOZ[at3Ѧu:g5R.*4BH1/jV7҆mVE[6wRsYbPRV.%G&*;YeGaMʦZIs vnN3:y]Fn/Z7nmK|񈝽+軵fGhg3}T VpmQyEypF<*Q)}F桭 |֡EP?MIݎDE6@tG_GM{5:R=0l> QCk3` m\^R+ZZJ?D`y`.V=m+;4_T<)(E_i"Ǡ[B4.R@wS!%H|Xx]&njﭦĂf<FDYra{tXh!8'q#-Apm> k A#Q{YBZ+= K&rBԤ"97 !:n:&&ǝK dI)8OjIA]jvKg26&yHdl )Z0'ɘ jdcʼ$xQOiCҎƦ5j!*1b>:L VV;XV`,ҊVZeؙ9#d*J.jЪ[pwS2LP3)ѰoHY?d2oQ՟m%s1M>܍'Eb #/++&pΛa T)"2%6 [Dhsb(?_u%,S. pG[g ί?0ERJi9@GVJ[E,P00R*|,A.&qZ1M\,kb`;#ZJ˱FSde1g:?fwٳs,娸$,{?4e;SK<ϻ86PnŴYP nA }7)d=Y6ĺ;a?\3~'@xeSj`w !i{/ǘRqrFzSqsEzs}yCG>ׇ`)gK}IeZjx}ސ}3]h-:'ۭ?'ww oī7UOg݇W2H1j@j^';fjw*ۥFvh{uP8JHDXHVl Z_=!l ) a߁@GA<KÔn vM>h,bx] a! (>Jk|eXn UG8l(Gp/R? "+x)bQ#*ZY-*)VJ֬VCioLM~d}h1i N0Zj 1'%-CTAI -J{ZO׀ ФMnxH;V׳ UʾHԢf}cixH^b4Zllhjm Zi1 g JZE3 k</>AX=֨~/u?:+k-{fF1~PVbϸK[+G+N?3.9"̈=i6wśo#[ȢK 7b DPEBZ'blzNȑ[Pi&̼T؋uJ KA[ȮLCH:dWF\x>&G\EL3KiU.;-9SkX ZU'wi8$T-֙΋=|aeɰW7hUYX(7TVs5pذHYg0k[ndgamm$ C/3~q_؝펾öC %IJmo}@IPZi[YYYU'M,[#y?V5g`ņec΋x6 #-ߔh"i 9d t(_׺9# s]R-E{C 7 %&-=puĢ">-S1^B:HI &Xa$t.!+Y:Bƈt~WEB]seb䰆0*¨J9籲ΧI<$ڨYd#Zv=!o'I(X'9$y4L|v ~9"Ԯ~x,f,a׫Q~'e_Wy@0xd΃ne r6\/oժ?8 $Nae )|sOoIRqyQV$k8\r}1FQ-ÌLVKxs`ʵP;ǩݙcH(:aŜHA dxWDAh ׌haR  ^0 SHIЅQD%1Bg P{H-SaN$fI* Fav~\R>gaNXEưD9({}+m.2o ahۓ]8!=_'7(Űtq2Ä(14^ ё_?]nC_,H|v)! /RJC{q7u[s IHs&H?}"' $N\sfנ:?Ϧ9l]X+$EUE@F(nI#塒nsf\ւ}kQ3)?pGxyU-V/?R /Ql;8n:V<\D@'Xv%) DnnHEݐBrA1Delgr9 =y*%FSHY YDŽM}JS+JvXJrqQ \jj3ydXA✠m]MD kNR/5)Q&L$B9cВ)YQ iF1b 鹦Ѝݚr}Lc}4wq 8όv\)r*g;MC Wc͍a*1V`%8aզ>Ir9)aTbRe붷j1q(f90dqA%šW€הp$'J[KL&RW >Bɘcb-GMNT!7X3XA6anc*ueG8?⑥ /? mod l5 \ϐ}F+٣e]`7K=x1?zO `GM]8izraL { b9eKJq3ZkFInmB 5,s8|" `x@k4 m|?L'1-j&\*%C'ݐDZ&"d&[jUM.uVqcm I)US1"ّ ~m ]~&~(u `Tl "4!@!%Uv2J%If92lBR%q* w"/uz_\{b7t pq8 p љf_^0T6Pں +4P[i?^EU.zr*gR%$ԟG&lmrEn~aV2ˤLK%(*ՆqFʥJQaөF&߁dIH5;Z9J . W4ML c+H Cf_nni7@tGhU.PEO8t%TS1*ʉ<J`6kAbS2a:rK=L}$z{ڙgtS#Q2!Z0abxHxxq2ecal<5.`k+)N,R"-OVag%`ly*2Z% Fv[fiSI1FB@Je44aMk6QAfa]h*"$A* QZLH{b+qK"'c978Sΰle˽ju0@^]=1 vhD!aO.~3p iݸ`.-y|ߏ>?3⫿ o%?1 =ʳε)%mk WsOwb1itJB5Z`_s|mVC b+ mǶudZ׈`FuWtWwB)V$v[`Q'6;YRoȸ8piع뎾VA (U\Wшtr"RQҫ`}Th1 ]G,4(m񎵅9_9!9eLj *+MLq…Kז lxhHmB}YP Xk(eB2OIѬ$&W"v>ѠEY e|rw L,2)RA(dj.W@n7 4dHAʟ|1&9޻57pXp}l>;`W^(襁(I+-J'$:.z <)+RL4AТ}^cA߱"ӌ _@[vhn&fX 6Dda)o愁q6Jؙ]8jBߞ!O'I2Ckn6 $w!q J:30ٳwZ d_K)ܷJԆQ`t4_>,`(5lȐF)jʽ4#{yru e\{! `)AR 8P [ vDRTHpnk^徺)*-T.5Q%_wrn}ϡHcM"]#,\NvWw|B})|ЈID#Kq.>۵ʠwd]ڒlhFfM~ӷWGZ6fn&Oc;xõA`bF #XH&N`W~>f$J[vyR8]Â.9bxЅ,G7ϩrAKՐr)BP3.v˵=wŠ~w;ar[uz$uT^ᰂvݩmme(x'Ncac.8G%K*"UnJoFQi,ICC}c*A@:ì .9&\"a,U#oI(s^K`%`D%UhƄG8.IӚ/5S׬َn]R_j5!Ԃ֟32nߨRy#%b/%k1WnQ6i}wS1oTn\g]D{nMX+76%9^nڸyIMuXw 7.-kX|ų3]}7;`P_(m/b3BN7_Dw$7ڪ j,a;su(] 2`\&)+?n9Iɵ|:P"B1kXdDA%fּ*)LQӪJrr֊+4$м]<")ò[DDp&T{! 9ElȄvCD ugße&n PnVuvML~@IW[Eo5 N;TH,f.~?!tV~ )'0~+Auؽllavk5gì䂺CG8 SEyO6=* .q%XiXcΘBc jxÕ상7|%5#8W#r+վZgz8_! P}C`@cJHLV]1vS=ܝLmAEig ɏt.QQP'Td"Dp-M7uP/2q=޵e :KJSt^Q IJ*@c+'4d@YE&sߖ҄kF$4H*1i![_ZY<pq 2Rx΅A3`KL`T:a @(F1Tcj0,VZ#9eu7f JL rW`VL ~|1wUKcׄfv8;F^#1lUouS*}+?O޿{Uyon?I2}_yHo 0_DXS\TzSgfoW%RsB懯wHQ$mv\HsBDZ_K)i/*1TKA?02LxTٮԈH^ooѣ?Zߊ6tIs<ڈ`U?bu=* DûG~$uw^ڕ.P82֑1(kP)c?aa$qf%BN::ǯo.'cqz9s=qLs:궐tBz- @jz {19|9;LeC#l:`F9k֘M%ٲA!:}^Xa;o4≶_Tr~v\ogxoVc>>3ׇpu;nW_^߹* ~Uvtb7Zt7tZbtX]Y_|uO_5[݀ٸO]TqD؝]VtFQ#Hgd.?/TK<{ e"nDѪlⲢZir[ݲUƅm,OgIIY5s)YX|uvLrWxKQ\\ We`fX:svCa^"wUis/;ZQ˰OlJ1&ƺiN p|Eo/z}ћ}z%RH(+AQbllH--K|;PVcXkҞ~8$lP%ĊP*J߉muHnO;Pq$i?z_X~F5%8Ea@ 1y}\N WY{R+I D|n).a>¹3!4S {1/(.=%߄LIUϞIf_ !˂R(T@~G̾ ;/2clb^/cViPrBsam}#uxC*g{?:j5@Anq ɂE7R@ vOkA@ueCΰlMwArt]}B*ՊY/ai3`D8xqSװF(I= ,8"_1>N`+$zo9},&M8fzLmص@ 0oxj۪9VgS֓ۖ k@[ǹ_꡸;{:eU29˵;5ށkD}[.9|# >8¤|)˴=wܧN>[UaQVwc$1Y18lFy]xkWҎW>?0:Q9L GELoҬ^pbD=?>*; .;㟒5J6aĘw>CCDKD&i q>B̼gzϬո#0zp鄴;!q" <Vg/z3,0{}^S9XW8TWM{;mV'3'"aYJ~xшF[E\ޥV0t֍xcdKI\88`R;̈́Èhg˒ڛd2PVB47c ] ͦ|=T!QF \H[ZYxI@R0t#$ Tsl)HUԨ\5U}G5t¯GD**Q+T ZWY_2%Zt+^h;:t_6v?[޾ jcIyЛr0_LG,ų<^Lѻ`*mOv {Q|Cr{1,j\fzt$!\DdJ[]&eƏr1H1hp-o=TօMObrs+$˖iOΖY>=8UL25Ԛ2_}{OHzE NjmO'CCŃT H$;qçA@Hg1Q'2و r+@6p牾EaI{WD'r'اah@H'MygF`fEZv25moJ0Qt~o(p0jfw!9ſp \=*z,w`P" Y+BS2:I0Gs1vB"_؆a't:\vp:δ:HǙIHU|7t@*#]0ŷB9Yp*o!5nYw_60^Z}r$5ҧX*P_\;ib{z\ZejMĊsu?Gcx-szWXv ZEt] c+'O&$֗y-y U@? 0$O}M})pF">G$1g&.I֝*ղR[6AH4SR '+'Ӡ?x8]2Z&a gy)k8(Tb&~jYY.)rhRlxZIO~օ\ xSSl{N!sŧ{_so-sD %RtɁz*zXUkuROS$2 9B*tRRj/%E0G8HPf0f^|$Bj8Ո8 k,4$ 2~8PDXt ״A`0[UdOTT"Z*6W R l="Er"XS49Eʷjr-a5'{j1EY-w#MV'}dIC~ȠZIDߍ M|z!͛^yHz8{ޒ5(hMgjXit`SP&*q&* EzZkѐkQ 5NܰL#Lӝw*_GhƳ;_6ݐ6{`}\Rbc,aisYZe10-6Qt_ 4XJ N%Z"TR 9`G鱣10yay$'qTRqi5348BP=Qn{su{V`;Q,.PEf4g׷ҵ6v*]fKndg)3fۗ`\an%K"$DL0X.?_4#Urߛׯ9. Ȁ`0 HN< p(^JZNj&o)`/ Jˬ#bAܱ kNsqBLP`{4΀UXN:^Fn|hD|]s^@Hu3Hk2eLD9y#R6` b*Lb$J}۠UKJ.L6;^?BPv- ]XXSV:]xV*KCYmua̖^P/wH(4!^n+1+ Nk,چP2r._L|V17:hZRMA (Wb ,-&0BHqֱP\KS֚\SILC+\y4xGe`5,j) lyW^6g[9@FCP^ P ;cyIbJ+$THVOr-f/TE(0Ąb .Q#΀[E03$m ' l)wb~+>s7&Tkw%5I8S e S\S m=Uc[.)rhRmD<ReKX˭ y"%SM^'i7k{[.);F6շvG&j.$䙋hLL.n#*$]Ž;J`_K*^l‹Mp%(0,Ckzmoqd,wԲbjF|zI`J!{CW;uK7%I|fʰLw){ɷw Co4TN(Y7QH֝ SZgՇUY2ΪQO嘭fք=@.+]Y7+ zg#quD?LȚ]y  /D,I:X-q!ETK $K|sUT*:yx** z£.[w/[ǟ(.v%l͎:t׻ϧ͋[{==dB?0Ӂx|Ms1_r_+/旃'l'>oL6j$|Ëlz? 6L!͋᫨BSΪ8*mmdlv,א={skLMSK~G'4it5OW%oI(|g[o3lb6; fLsCdƞ5d;ܔ퐡koPy/X( & 9AcvDAx=NOZ q*&~OZjliN*lTJSA6pZh!TxnD~ێD%r1$~y779i&+aԀX2])(z;?lO#sefɕjQpReT2xʨH Q.aP..N*P%xGY\mQt'=4ד[ JMS%V-QYPQPJTS`xR . _~r> Vmx|zTG19#®GWKNQY93IJr Y#,wg'qʧ.MQ))h!琤3%X-iCEfy^E]صaܖ$w?$>oHهCCyM組´*7âB 1MH>ebgO~:97ϗ)9&H@.\R뛦n+ovmTKy1la@A|@ڟˤ'JtQC?7;ƎB~Ň^Y?; oW = tix}FZICM ;٫>_˂xyO6NgӿAK@]<\rUd&>G% fJyݶj:j;hݏgrhu0Xd[F\3Ӛ]AJ3C"'蝟;;+uAnǯ}h2P3R[l |K3T~B3]5€@Gy/81V'ngڞusnRL7BRxY,D@L,8|S+S P Fgi\.E6 *M5*mU\ fu5j8]_ ]~eTBJQ2dqdsS~7kOC 0PWO1~X|+:Rh%Of}|( 1EWPɭǷ>\ a2!I6pWÇowwV `LhO)~ar8]»QN~ S9g'3+ Ս+uscS;^|ӋIzzEn.uz\0%{iEm )6)O5tX΀_a$c)FBhv3qI`qp4Gc\v4V->!4#eX3) u{\ڬ!OAX!ρ_? -\?/CT&f]VT,ϯvIIV[F^ϡ#Tcdv? |6q- P"7EI&٢qѓGϻ'* _ǒƠ3x5cS2?ec;SjAPHw8;0 }@ ~q0n*Q% <>S%⢝yf89ұFĤQ2 :̌sVQ&,h*!F#M)) *^𨜞.pNOLGףY1FDZG)KR+d+*W'qM8lOW}vg>A_wu0B3FW/#^(EjD7z`tX56ZK p}lcZEAS<;ݥD#ꃗbW 7~:+̹:z86;b)FS-t(֒w(Z<:<i8SoTkf!$K5ID=>+0Ioi꾨Aݗ=r1 EU ׳Ue6nz5lNk_T/VZfȢ.|,*}dQ:}We?Z4 W _NnH4 &o')grb5ˊanXZUJ8B)6J,䝠BfS#6[L hRZJ -x[) Y2nRK}xZrDw 5M #yƞo FWL7}a'xEF*ISo`дSڷ/tN(lwew_7EN«} v22C-"Q#e94ar2Û<tpQghWEuXX}h uguScfgY8 wzeXKPӘ4% 3f%Fʬ-ϵ=1SwiSxa2JxnI)&_̙Ȍc&w |wjW;ɦsx*H:W6\PgBBI&N3ApςALggM еMH)ї᠒8g9ϟ5Ŵ﬩J U\cZhi^5Nisoe9㎂*X5i툲Q9NB]JpJR6\т70͢/(Qz Gpˣ+9~@q나 !H5uYqGZ( ԘbZkE[0ix 8M/*񂞀ߢ&Vd3eDb8 %q ;Ma1lH7A0$fEհjWAN&Bs5R._[T>4RP:`l)uA^D ݣeZ{e-sJ%L8]b T ]V74*/7ag18'|z>]^\ Q.RprpNNx"$y@$)f(L?lr?v~~v2 OgD[e,B~6˥mq2p>I?<7UlQ$8zryRKӫfFeJǖd Mq,Ղe !!_)"9_M:}.Sq7hoJXub^ƪ&Êԙ4l*'ai?W.ՓoƜm9M,~u[5yYKwOAv8Nu0)kȯ/r+ #gt+IJu.!;6K+xS1 'LbAԉČtIG,1tpaڤm51)3'32\`vPƴoP"_17QMj|JY#7JHcjOnQgqƟ3i?ȡ!Ԙ4wgqjt2-)%v$؃F\d@0ҝzd3[C]ꐾߓh|ZɎ>LP*iNMhU"`:ՏmvIrq3׹𒛌<# u2:%(f`F B ޼):s% (#5B| t֨X"Qt%)LQUQ't4xKB*y<J%J.W^5 wmmX~YbyX$;fe۾-9l.RKql\$<$%NXi<0x*P<3y5M Q0 e(y,M?5Hⵋ?->x/o(kÏM|FY'EmQ QPDsJ X,QA&10kTT_ޡ0k֯5.k2%H1u.elap7D;P z \(d ¨q1f* I-Wa-=ʢ6+;5ŋAYt|ʜ^1m(\&Խ8@aǿ{sK[.·1 tUOvWb8;WO_R>-rLUҶ)Y}t.HLR]Q82pTCe*?{o*,JBJY#p8+ɫx]g{nqjhg-Fvdq?ٗ?=[]Eo"m[YQuwW$߷9]p$Sl=i q(p꬐!9txz;)@Xq6\$ rwreX>T*^ޘ\ޟ-3vww_/)CoN[%w,b7+ʹ|zYNt4uxc ΕwŇ|pO'tz>:Gk6%)y]pA1?]j\HymÊyk F|̮Gq0Bdއn)˷]X/BʽiNWxicOUW!,~I>QX֊RJ)K=y'UZT|g=iPaʵT_Yv 1 u-wJ*CD0Zf948#scC)ǥӰwFmr~RaryW/q<,ؕuw}1m#&lq~G2Qv#?,F8;ZƢAoc0~; &9-MAqx_tyi1!|zC`M~ ̕:k3J-ǿN|]vEB?# 9sM)nR-&ZFv^T-[rCz.,M`bA ܙqAMjx<X18A 5Q+:/_jA]G#H8L׈9:Oګq1FD~ Hfi'?Ax ?1mq e%anuխ(VMQH$x7O>݈iszEfs-ݕS^hܟ(B\#%u6W~ 7¿N3}(D_f>h#9u)EN`Bh՟' Myf+'_Q:y XЎ^WP./s-]*\)kcDp_cJY_<(mQK3Y?;t唑Xu-cs N템"9䗥 Ϭkq l(Q]޽a,$BA` 7}FN)R[irnf4";㽩ie}lQTkz#x]]OwQQeه7^T0/-&kfM^Q(m͆vjuA!I}X#+4󖫺v"""$u]{Pۡ!bK3÷l޹1H.잦ә{0.!x၈[uZ2Kf* KF*ĬcR+0 bZZ>50RΗGK BztYİJ6|DGfu=Իeޕqڈ (l%w;Ýn8D@]}T[1aLHp5-v>ʭ$%1+tFWm2<%!R? oK?m ^sWV~wʅan))C[WqV~wʅwřhșx0u8HD?ۖYQPY!aнmfl#]V(;rBA*$߷Ar)>G $r*f t~v~Yxx E]-M[5Qj7lUv;9+vpPTJnLeNstO'Y1V17[3PB $L[{ -Gz .V 9a{5k *冯Ry)i t)T`ń`&w(AU'*-tוaW,oIםs{伐V"8"Vh TxwpS` *ٖ0\|Gk*q oe': VZٰxPQSKJpR By\CӊO\ß>am#d-F ~zC}ܬ VLop7rٲ{%-Xћ{w?H 4~~CNuAB)y,n<ޖS&pZ?> at++R>~`(׽FrBfU_J=Tq^xO0 ˻I^=ƛUjn6~&yX]ޏ)><,xijgihƛ`hlc)˰pk)-@RuY!+ݾjtV&BSgQK6B];5)N HX.i(k;b!a>p}Fl}TWQp! ˞3ylQ4x׶1z>nz8 R8ox_g:+8Ez8 dc]qgjPWgK1Pa[6g3!,~6;bҥ)elHB籢|fH!iZ9΁ Al|<8u3 DDIuBtɈ3 9T =X k>aP {&MA㞋S9M*9GBY.z.'L"}>qzN7~1aE/HE*p|SAT(Ai^O\B*/?Q^)Y)Od ٮ~~?vvr?J5V VSA(= ՟~y~.j}\9LVS&XrfƎ,,U$l%/ؤ^շ#IxfK_PtOV{4(5ӜLHzC̳_Ϟ1TQʼny jr;_ .L1//=y*xCzfW M %hڗ!B*Ը P{O .ZLeۜpq2íF7. 95\殇0 R1KO8 9AL`<P5ZOYn]桻&D`e\@dpr);WQ%I@+6{M̗Ϳ?<[4| 'Ln{q\U e) \ƩΜϪ' /S8S8)O& IE6BPt<(dYm8"L!qT4,$6YH,!1A,U `NT)ԳjPUZ9]67D~htW) y* ͌3$Ta2d{:zMVBJ70[[Մqļ(%iI JT~SV#(֝VrûqZJPvJֳK7 f14׍-]?'3 o4SBW kG10%GTsRh=H00J' ObJISxq܇a Z0cͰ 7 LCOc&x (F]EMКj9<}wkUgJJv)9iJi*,1G9oN@A ,eDaagvnc=ZW{vPIb/zVr7LpsT(S#O5HL!rTϽEJT?b3d<ƌgmp!.Wc~fy޵q+ٿ2ЗnG|?,.nrmq>%0M5Fp,߷8=ZаR$1"MOxN"UOuM~@-& #.c .~|{ k>$on??f-7x fځd3HMW:|w? lb2n=Nf}״Zw#=H8egY~ڥGާ [> g-׳w &F1p|V/Ԕ70Ef/`?tt#,$9:{ω 'O|B69_'C\(RV%+-L^*`@c͝HVHB)8zZI\f3)Yk#f)Wr!efc t++C\cc s0X>&0veH04$mX=04Nx$0Equ`1SN*_|rT Z.Vʙk-\5%' 7RmG zCވ]} XUYϓju -řtp0^֫zu_Y(o3FIL .ڬ&lgy W01."=BxM\3~ D<3"lP_}Jcpƻ u~X=\.3od>Y6/= nRaٌ>ˍI}[[D}-&ͫ]㢣6 1+D;3a-ՈzIuԦ'eZCtvbӢ)3J!D Oef(1l&Y.q&Sҫ(r r25I3A2'G,#[fPni-;P1;J"ϳ\.,BEfח5\A\w~:|2%V&f'W?yU ~i*_Qv6AGP$(Ǥ\)0V]|ܡydR5QK:\ӭ>CQ||X;`Vς_S)j ى!4hu&*vYў LY9^f׍k\93~i~ُŅjH}scD9ɍ*$k_3ZO:m'éw@pIO8mD2_"}rرyHcҸ{u .U^&[6 p齃V֠ݝ8gj9`>ƀ!JSPF+tx^E懹 G@װ}4nzDIR}MوS/Q7 Z1 h%93FVV穓is33 gX!%ɝ$J4W-z\ȅEINFf &C8| RG)*80~ȑ!_DK<# 7ҜoJGm[莆Z&4䙫hmI򩈧\nzמ|t_mGLl[ YO'sym;\7:cǁc)/! qر!N.͉c! pp%CX8A`%, ,#qV =4~DZrYWfUF/QLuV>L|(>7'?"od?ЫT[x*"a{Zix.94r3FDZQ|I/Ԭ,iTRyQLQۧGv]ܟ'92:!ʾW;K0Ȱ Zo;Ȥ 1hKm>R#)IKn*-U^, uv I_'nZ0djq݂I_0ip;6-rЖQ$Uw0{ph(!FCUUުֵ~ik%mv"vh)Uw{"@EJ>Dx.R"Bz0#%Oi۹q%\4IIXV&qtt(*4}t4Q}斸aLz*D,ww"ZnY,MO]HJK{ҳ )MVJʩGRjC Cf+jB+˶q*TF}5?!Of^I}Crߑq:tٗ(ye 0ރko|٣,\S_1 Y\bfI:ury/_Ώ_(%GAvFȭWGUy7ߟ>kdyØ`uN|'J1g-Q_i4,6;b"D0~':]RMڢrj&Jyb6s.lS혔-mRf4s"%ΉȰ B,6&¸=7r#ŗr7P^b/_i)ھƐIa8Cj$v˧Fw k SPĭ0{h <xfD^zOak h Jjg]Xfm@/`΄j /8E^tO yJvGaHGovUM A%6O׏!_!geƌ0U^-(NFh|PW?㧴96sm.,`wn,g_6?.fٟ3ip%DK|㚇J(2`h0Oy!| zflJ|y:9酈>A U2X =BtHSG$Zİ*7q&yo@B%p)Kh H`8.WK~rF$l[`=`υ! ڪ@G{bg5fBl64ӵXQerynq32Ь̭;Kj2'b\2-W,m"*Uтt>N|^kR&DDY #JP˰Ԯ _ٕuE8GB/7؜nI4A@R{%KFqx҇B]j{(FsϨ5 EՆZ}5>z-g*7 [(쫅IwY5Un,jɉHg/E͖*ҿ)=̬5rf=OG\qE̹șܦ>nfwlvwzrBq#'^.s۳peG37ZQFD?{CN8}SژUhoïg,q<3sw;֟-lȀlc64ruRX9M%,e:eu .w/>yAqp=7CɌ!ea䴟'Eae}ae;F>|ٷ$ 7*&dxR=Đ3C.DjԅPv1zk!2D?)a.^T7do-*He E7UNζ ܫդXPs33z2zj?FژFڧ}fG\lϯ>SD/$-FmeCKWrͺWҮrK.5+ʥ{唄޻>4t+sT6ÁQN!iRgir4'D,*NK;+"[Kt73z(0ᱚRauձgWNj˽Z1z 7#ୋdH:jSmhWd7[VjE7:tm+SŸf '6{r Jh*` *P, KAܸ'(!bP%+`7QpX7#Pp+KIp-z#K)!4lI5-@s3r hvr4Ӎlз)@s&3j%ǭP|Fg{Rrܲz0 &|c=G&-F W|@ҮWe_'}öN&JwyW%+%VA:=c NAY"&!"d˥ 3P 9 #qTiGExԄ*L1E(LLF-M\mS G7AږG9AClSɕ;h9&;tf)Ud&1T0] gE!p1@)L p7tMg"ʽic/HG hy(U')Ɖn!N7f1܉rۈ;̈s6vtK&hGRҴ"[%[`iZ Xϛ m{v$C W}6QUC$( >Ɖh kӃjxQ浂tx2T w;hbv's"nJŒ4 S$5w~qF18ha$Cɾ Gss[%9,}nf*9Ë4vb+!r*ycq" ){NH {/ÌdRP3>Y;Buts"kr)K$4i2韕YL(O*_d5j|Ƌ'P@sTI-2j;:C-SKQ~}ЌQ߆j/elrt:|E? pJ5!kԭ!U_>~wY+#+49󠦧&?Lϒ~ۇ C' ezB=m2ShyDnMdaCzWr׳L}fq]=T'\mlȀFg˶Y51RX99*a,ӱb@UhW^>yAqpXHf|,BBԲP ߊɒi~Ta(R~\OOI|"SR)cP{96k]P}v4ʤgT>zE~^uQ+U2 Kz5$ '+%xb!JE]7S8^2@}>qC.2S+j('rFv5YMі5WkX0c>깟)Beq,\~IRFpE0OqXM?]9{qnAҏPI0Oki~uoM)EKuH~&G,^ٳ5b/^U-у?8~[waH&R\%wM-Y`3tTo8sϕ<`>JL}\lI%GNR#z``P/Bs YpLHƍP-9Lr_Pm9?FKFwRdAygZ&:J %1aUmxǫ7lTնu}׏2MF@=b;׏I6r~-M -)] Dyۗ.|ϋY?/~^4YP[eF1ApZ!MNNL9Uqd&+.]EJv} >ꆌ}z9f7 /(DC짲2wI, z'a൵QFp#@0!\Jh Q)6 e`>Uijڕz$ (&r"17SтЃ7xju`fL,.n0(nPf 7KqV!ދt)niasiD7<.ӍmSQLöﻸݥz 9+n`UHֆ!f;T3>[{Z*G 2@%RC ceKp3pYχ`̔Q1m'TN>)`_߼Qqb3'ι>0@yF=w~Pk|\|qXSn^N&8,E1#k_*-PXN7wVXꅧnLnr F f5K6mR_;5BJOW 9Ѻ6K]4F^Ssvy`ǵQ,`گP?$ްg $- T$iytڍ;.=5&#p*i|Z4%_f1/Eiux&w4ʥOg7C$xK1ڋmEcYYG.>SYGbrnnP:F@CA٠:+W8ِSxb_Rnd>{^itpӣ}]|g85<q;-1;+eR,C\e'Ul5!I.Mg3:eXafV英9XiI.Eɼd?Zr*GGvW7 3zd陕oUS]QE+:)5FPQct:gDk":tݮΖVQo8ek1F3\zRjOJ-R}Rjo[sW-y[aϞ}m'r"Ŕ xh,b1[$ZsHFzϽ"]쫔(`Y#}oU ȈJ՘s:)r~ɏ_6H `k*.U$ZbHY|/fVMb:mr Q|\R)d.B*@!L FED&m==g(/v[k'/BJ91 qYHerP+|ǐZ .s~p!&fp~<;78?bǰf9~gSa>$ee`~#)!7rU":^vu!9\y-Ya[>y"57.ZU`j I\=| ەX.xek!`%;9dC#aiݐ*z~ Bn(X]ndr9<^v:ŝmܸ;ȡ +xmkGu.nOmz"bf-XWuC&Zu^N,YfJtȭL$שw]D=mA {^sҮ;D\/7sWt1?w`$ޟowzY-s4Q/@>7}6,#6]'qJJ/Q5!F q5 ED'c7 9GlDti!^ y,?`k;Kb-_fUW@g+ @ޡfl5@iH98!Zkz̵<"Ya|h20| q| dC1'~@@V-k~*]U2 ߳u'Z!lLDSgp`ܺ`gLy }37UG6 (]?Cm@(N)GAir<8z߱kY;, >~YPpg9HM< O Ea(존^|#@_ŽC5~wT?fLAyFmcq)ds]^K~1+9 Q *+zxΗL5~< 1(/<L#K48]?&mALv3x֯>pt >Poӹy'<%cdǸ ۞Nĉ0=ӷX ݸѓXOk6/3t3% w[D%Ay ;޳ŽK+%;z=&ni hAP,gGq)mC-ʍOH_R]ڶ qw =U81 (g+;)0+H3Z U+wy!\6g|hBOT(nDiUaMqȄ+Q//LPc;8C0/ /NyVyF1@LoiJ@HnwtaHyT-hxf$m 1{g뜫qN60E,z +́<㙑Ǡj$Q'aLuYaQY57]Jm Q~9 VNgu0NycͫCTU#CbctAY4=+́qffΕܠ`gE ^rq|-PAc]wơDw24 G(9 : #5) r0diWUe4(!s29blQc>ѩ:-{1"r^uK6TJb2JΧ Z讋hH B$ Pm65k*!fqՕ&k \1V(JP$< "PjTS Pӽn:]9֙J,c%ݺB9{'LMEP;uT6X F ShaMA,Gg>a wmF@!d8@i,zh$Yj^,o[]zsk4` n,VjܑO7A|fe'bfJ72}. \6.u8hyR B7Nӿ?G|DPhpj˯P6g'.T5GPc]E);0l"N3-#rsyΩ^;) lN-(j{K%a~-cQd( 5+5=HDgqݐHDTxAY-Aڋ`XآZL1Bma19؝4#ES)K#mƣ8D[/88{9bs/ӷzm-^ídiFym/8c!Omsb!k@=$=-Gh\0ƺZ h,%?JiQ 8^Ci<-@\ l 3Xn4)̑׶ΟVع)ӓM]nϏvԷ Ni(M:u0JBL$G0Y3F`UN7T7!j_3W[SS- Й(vHN^_$S*%IpI¶U{jX 6h<"^Y]IhVOjlMn%ӵŘz&>_^׾O~{ !И>d# y`rSc铙YTnFx_Pf"sgӒ iӼVuPt4ȨB2Ž@7ꙙۓs!c $$,}v1ߘjR$l_qb(&EXR("qPdeVBZ8D` L4j?zծQ@]J!<6'yjG˞Ta"EEYJ3*XBŃ/ W9 8l 6=jNpU o z QĦUv!YgR #@=mY| l㱈;-0s"5 (`,[VY 'PMđHYgTT Xɮ8Ko91΋X\ٌdP0SFeH*NR! qr xZra$HZs>#N 㨆,Uv~! "? EX]{1LBsf$=PmC͇ /Y=n8wx̒Ž/6UO7dD %t4'D9#=ϼFdKGd UMd~S{#ѓr]lI8N.OsA߇p1 mͽ!;wr-4U\O!0('+߻$µE9}W%'*G9!4V({#rFkEw, Zo}lZ8/8ѭ5tqg'J⣓Q*ڎ7Ǹq+ul;j,9=1BGhq؈A,˻j>Uw=TI-s6i]uH(:}ٽiY dy! o;+Ң5ro/\eL"'0BoO5;ఖ[d̂7H>_=>'4AATZ[Xqf? ׷y Ma(&S[A`| *"thi< Irn HpU&3*ޜ*iTEΡ/ 0>se'oWmK֖C4]5pg] 08&n_\-`QFֻkIelcuΘܼ@e[ "6e[bl_>14'ۧoP6aoz5aߒ1S>D{qqO˂ӮsWyJa =cq{. &ƹ6.*lq1)$YK؇l$,di -ṱ߬gp[S6AQzՀ3@S$e^a8pBAH=kτaWN)Jjjz$nF(AKb'0EkS`(/ّa@!4F6}Lj_^2NL tnr.і;y"ıܻB_UE68Q4ѼvM:m:l0)ǨJi_rڪa(} Ѽ]GjÂ`rwvXp\Ḛq59zKYzoA.-H^[%.<d'x>€_4ė6z֋%kt U:dAeYkh;@#tY4`ި\7e,>)3)B@: d]1: ruRpKB+/j#'/Ԇ!AHB_eHJhD4:F' U6dJtk~qk/>5%FŃ+jØ^g7~Mj )z!$ZĘOOEW5~B5jղ+FT_rCIq~@J!aRi0 L+Jҽ9)4h3uy&&z >)DrLi:#IKjg- iXilcM=lj.DS;. a`D%R*69ye5>x8CGȐ7=PZnE/H=|5_w# z0OO`L")մ 't&-.iT#Y \UҤ`|gH ׋OrNP6NM͒[ɹueL-y?Y1w\??zlph5I) PtPx`K_.' <e]U!wޠfXLWX@V7z1})L޸>yŀ@|hiO4skĮ8$A*QfG9#/WYY4B=+0$/\ Nڹ]kTN{\SO_8sK"o1Y|!F0^wtɏzpW\,E+i]()u!-(|1V3{+{֢ tt@ ڷb\fA YМ=͌f%~VZ[Czv,lo܇Nx" ‡/?Oj5a>vbә]n5N>}gHU}UU l֍zi`dF;Nyv (B$}p ~ie,nh' sA5)0;P#S LrF%cϻ/qnPb`:3:]pq-:Vd$ ])En#m2zuXk`>5BT9?^x&Qts2)!qQy.9\ogU;[^3픴BNIRBg˰|,["f o %g(HKo5ꛖ_~W"<4o@gGl Ƴ0FBq;"AekV.U gkvV`+SȼuOOuhyFG—(aH^ C%k}ɟ$v;>_]I-:H?Jﵗk/^zһyYεj] *k0QqR2A5J5r_?<&y\m|L~j Cin9n^Gn&OmN GG5on g72-G [TيYŨE eWRKj [;NZ$ 5j)pTJCK<`I L  !<Q 5p.%~|X- PʝGє.|wZ Aۤt $f#) {T*|; O\$0Ar;c:{N \W|T2W/o>}uN%?>#DFֵXuBsձn&թּ%4/ O?a""DюdyzDMi`Ed0CfzRA3y' gUSִ9΃)*"K+jf>ABZ:e x=Wqw0A`eB[0wqηNcԯ`BdjT}aH[ [q3jΗRnF"p虫o]@kpƣ("J 8Ca3'P}u"!.XYS 8.6jdTe*f 7/!#L%Seo0E$hEK0DӒk9T:jf p2ƹ6/T$%Y C "P_x'R %O3te,ᩒ唻c} gW @|ci /6ҰwK0ڽAg7LN WFKd](}t-uo˪&!M*QEܶa y%96Ze4D?TSfOrbkZN-UDIͱP*Զ㔩JtTMRHKleY:+M6JjI%q8'bb˪T6@mng)Z $]Jst;郥9Q,5(YI@ߘV^w:@7')-8Vͷ$)^:v✁ :tqİ`E|l %"թ!#ԩq͏ ćS;]8OodMLáujk2թ *NR4}?EK08+tOBU18&y 0v |n%?O)z&2 Llc܅`v.4{Lzu5z p+6@=_{iCjǷtBQp p>i>:$RCcptY0[QAurL\urb(JD ^@IX-˕#p/9SF;lom *i1%HV0"y"V<N<[1b~ n18ӄHG$oWqUK[+J 3_>Nqt@~~+e=x{^E`"߅2/vu߷~Cky{8 U8(?_[oU%F7+0\COW|%ͷYG{RܣO~7yHW5e>蠼w3 _~@枏@Fo,ыS}TX5`=F)0j7zGgQm:vw):T,3ªp;)`R2aIJ/<&e2%}^cޙ e xizX_x9yqٵcٙΏAdD&q(Ndc4a`<irR='\q qx;03Jdjwv PBI2sRǣLPϳ.<^1tJB``aSވL7vl Q%ui$e" RU iC8rQYJTYr&+Eh-M⁌SLVKebFZسT(\׆[,fD+D * `03+ j cAY=<` 0P50I0PJqNGle|7ڝɐ ~&&,t{P(}>]ӽSL(WLN\#̈}rJ!Q07bpwKS&I8ѽqam_?? 7p{LG)$[~ltz-yC|[6;{-D<_|i7I5kki~m *U0~qBX'W_^\:rKː7\n%L  M1cXS6nf }W,uF]'+ U$gŬ"@(i '/./^ꢚԿ-VSw{=o1ڟ׵3cC:5U'^ ?3sUt?̆tgD;-F׍ &(Z; ;q]ƣ+}-9s> *Ɂ˞CQ#6'a 9!8px0qq`Be{-ygTXZe/zkmG7\I#荽1]:Sm[|^{y-7l^(\kZ@阃uI ^R0UJ#k2ѯ WWża_m͌tO/_N|gr/c9an|s6[LfU珟,Xx8X@oBtVHث+^?4tN~ʸnΛi![o\|ziX nl'Ja@(bDc@0ER2`UD۲8T)-E}~^ S4kQgv d" DS)geR$EXքaZ,g褛}:jތ SA^dvmģc,1Cܢ!9gJ N‰rbluvl3Qkx7nZ8ٝO&r<鼺NҊdw/EZ)]襲<\%eu!̧HRԅm45.WHUXcZo8p̦SZkv.ܘݹ9ov]~viE; OS;w,h?z/&9c =XZnv2_}_}p.ʝ'7՗KLwJ.o]tqѷ2~il7K5E9ЁɌ'Y^݄d.돍 xt%hD 'NDĝ\D؂TZTgyCvy s14X@Da vh u؃v:=8 o$ j۠J%$Wr 2*G3ԛOJ*zӀ;r[aPsd9>v܁yiȒ\V!G Ss|.@YS,yL9sgbx2/"ʫJgz~ ߲O) *GC;d sĂ"D5Sɏ}g$jvI zP伔f\(('~GGZVoqt㐟G]~0 (Gp )JX89D98oB&w`C?q[I/y vesMaA"oN߷(۲nS${Ɍ-KWbXճ} آ.~_aղ1+~hpkA xG+_,z_?zA<=Мބ'}|KPozsټS0T 1,7W.ͣCyW\_S^pvBr 0:EMdԆs%R N^A5u WZ~)8={@_EML{IT ew5OLګ) p|1򻹹ses_lpbc>J;#w>Z𷏝{ziC>qSo>'bDjO7 >҆MEEK߶Iz{o;sw-D'u}y|ȅ{j\Lzb(ӰS7a}j7]+M A0ƣY]7n}~p) ͰU +LlXqg|x*Jc+oロJ) 7B٧} =7 WA E:1L |iŕy6ۭ5,ؙʝ` Y&sՈ:DYnz첄Q+iQ`DY&qZrH VuF3%2wZv8B0gJFo?cg|![sv6FM`Z՗-.q<=?L0L* ^9#M!VJP&A5 );]ZvtJբDt/_P'`g_TZtt}a$A)lBH*gt}q'bgTS3ɳCaQ'5 VZJkߗV1-Nю#-9Ŋi] xD`SFxb"ksMtbs並7F`pL[ >5roعS5.lVh.e|"NnHŠ턙%LK.Ñ/Of!폜R#P\! fD4u$o ).֧s֓L$'ښ'" j=ٹz?)H[n!0&[gGmƯ_leË<=~%x咤"$vL?yWgؒ&=-`{b# ]qy|G2[/ ;[VdT &tm)p^cpI҂NBJۅ l~4*hU쫠WUͮR+qν'kU5sl(BK $s>((,U\iF+8RrBg8ˬJH3P(7)i䰲D& !F0L0Ƭ"QeV[-jD/4 #PLX"ȱRYc`K0AJ\!bg٠che@kC QdwuFKa uNK8QA7zo`[0;q.q}:_y,;W3O] ĚS ³'ooٿn<@_pm)m\K|.tw?08);07._LDӮ%ؒUmωmݳeeƪpP|]8QiUbN{iY ӛ>jvHGNagh"t%vGO+u]4PM]p elYؔ ؿ*'_/˿_vqN>9|1 )<}N2$$ZI\FyY9ERބvqsй#74HB)Mʬd^ m1q v|,.b苀E.nDҡl-$/ jū8dB %wkX$\*s jjV> fJiR0Hp\4HB|?ɜ*OCE@<o@2ʝR(KFT+,$衎+PgRR6"\*,W:bBHNOM°$"BjP=jF@')]T%//gd-{;ج$@[Sk=\#dwe3K БS@v >< sź[j;.{<0-_Y:kvQx$lGG7!En_ccZ Tdqף}nmW{3`in7TuRpT]}>x֫Kյ޼,8@c{r&phUe\k#XYh:( ݰĿב;̦ r {r#> aщ˘0ʩp8BRzɁglrjX95yQ*h#.{DQ:HEĐ4K_pd` =:^彮 2<^w{]W&{4b60%6&N/F(GTe|0N @vM-TJxj\` h@\fL=*Ìcr&1=zo+'tzS<2@Fx>ue@3EfX2{nyz3[:Fwb%N" {(V \ iU! j~]ՔM@ޭCTx N`UA^~^/zQ_:QG&DcRه (P?Y,P̀yqZ*񇟤jB [%fmmʒBXQ9Q+@I7i12#I pFa-.S4e9(ʲ4PuV2a+$ lAigpfҖql:eLHSG(픍,[;r(lhB7I&$RYĘ2*ވ(xz$^? c%?0EbU2ʹ83&l 2K2TI54^$f "įFRbs"- @$xV4-–rƱ<`#NMfe$r$ҕ[F-u⇍Z=z Fm5L!uZe`)e9 kA% j}2G{.2i` G;U&d?쾜l2 9ټ?㈉؜"sfNze n2S8ZZl0\1d ܽB5HfeTinz*`Q(xั1l'P09AKIE_I$AS(Cڼ]jU1;:;|lr},i Ҥ4K5ES+ 5Yߓ-rNq0k=vu V2f;i AkꚸejM]u!ͣyq:҂ݖ`57DսKq0Z" "tOͅliŹLMppz݁b,poe凱dmaX =q`՛&/YmVzJڶJfro>j'`UZ'$Gcb)Bǐn#UXj\rfTT];".\1RT\)Z3%~x;^ٲgLp@3ގvmܕƹG2q'A GؐK 7w~Y|Nr5&TEg>FI-)n=T=V~UeyR֒|*Sxj8CnĨN3XvAZ[:dukCCrM)q>ni-I}Fv]ӕF-y֭ UtGHNiŬe s1w8@N" I($@–`c+#k nH`; [g " i`[g3L*J 9GWؒJ "Xde}>s:u&~ƣ.Lu.3\ {<gAA}1B~Rt Z)VB3@$̢zR./`F;:w)F[U2/1?)vWC~XznU͜zu$ku+&_XJRXkZL< [ + Eh&T ۦy! REx!$pC65f"!:5<2NOi>j$$vҔ4E9ziU%OF-*)1DI|Fd SYJb[هR8G@UZ7/tn$)*sW9(d~$=O z@uFΝc0H&1 xf)!mq Jx~{Gs~)u3CQ{o1p0 &P,|Hzo>|N}vlLlŪ휣iT ,J*v@XǯkTLUev@R 0uBdi+ApYr: ې(D;BA-ܙYR8< b"2Ȭ/1BqsQBt$-5Et1x1Y?{MP/ A7Y0?YjW,yCUňDͿ~W:z>EvX^xG6ң ,? 0YlB;!XhO* *h]0y0_Ĥ?D6Q}驭@:ߏ=:(F>}x (Z m0!X\wVf'O<3q<]oPDNx J8NiScc:iYpw ex!eex]X@ECM_ P/YjTT6"G 'O.a(5EOԈfŀ~KExUB{AlM;NF$7bK^`q0}͓AG8@f3'7NYvSY?Godq}xkWUvQاNg6u0j֭jm1LQ5 msLG:Z{| Ks2qzIDn $_ŰrDeiPKԗu< !zYM5%| 9nxQʺ " A Z.(:zRd7٧xu!] 5m{񘛃K}N4D'  .tVH_/V3@RQc䔣.Vh`1N1 t&D7޷p&Z )nDC*p\bd0"AJ2+҆,I6Ds([g?lX B iˉHڔ'!IcX :N-("[AFوWq[ZKȗ̿ӽd/,J _ 1UwcW$K6Fz3ٯ~[{!ߑ뛫^:Lg >̇xPBݻw0H~;B9811ӫȍ5C9vkoG}Stg}KT5?!īnhf'A:pNٟ^-_X`/%8)窱í\) ?_/` fFQĀȈ%G&VPMcc yϦSۆz9XvrO vT H{9.)Y B1XTsYYbGe^cTe ?⮩p4ΑR_։rmIڤ@D*;&=3^ ƫ4g Q.kH… q$ac UPR$֩ԢoVNK@ݱkQ+K*2?KngR BR1M)7$1 š()HD=u1W((KcF+:}+ĭDDqA(dfn;JRd|3: LHD¬H_}\̪ LZ/XxMS]qv_pec:'ШzdaV]-/>NK\ΜlS3ӄF,:R$cn۴ "RpZ]]0: #a%I?2Ay42%/ny4A  b< A)3TÈKS )JyǴA4Bm]V+α4 C D'@_bi : \He`82V2TX̨m)04a10KDooii_ aɏIt(SCRAKRuz)"6\YU3B5ToWFJ߭"(»k 3&dxM'!#˔"BXM]fR xH gr$ba5u9rv'1P5uOꑉwנ7SBl:,PYWY˚d6T9uS+%G$$DΉUX^Ξ:cLZJl,UŹixyRUȚ8bNl*]{>}}5Y=}!R<ߧ1hx 1j`ȁnͽ*lY!5bŇmO9&k~O/AGɨ?z)D N26K G0")w儨h5+sW9:@VdO*G*Qdd,A0s=e?oM Cy^mﷹfVS0ezx $EI##81jpO;cQOzk/P/:c:K(U-3X~V9zACB@;qޮ|At*J܂{'{gҘ:b Hc@# NWmV zxu}M!1@ HXT'"V.6%9DLrF?d|&'Kݙ6e+XOIv&~j%_0=`91c(5nݗo; ayGE1p&H;tPeS7zb[E.V>xx ZxǟN&QMFM'^,7kMڟ)Pru\̰@A iaiN%TܟTw()%~M yBjXi Ae =_vTˬ:A$`_EXOxJyLcĔeBWa4djٶ˧nOTmNBO% M0VNFQHsD!h5ORc&BA L"|-/\JrS'`kJĒH:m`L0 e/Ӹ%u~ۤ# %g)boK~m--׿,ֽl9ړ>Y=epFӻוxE99,8",EY"[c˲v: 6ث|6_?,U0p~zqu܌[ !*Y~rkYr"P8 Pփv?q C$\BoșLFO%73=95]TK{{oWECy]{Ӎuz4>6$!}&U@~Tab, ku9GLUġyw^̇uP(0ʳ )\#V$% s~NY+Fx5mbRvໂ7y+\IP{4 d}< ,l:~}:? s^]_-XPkxG6ң ,< 0Y:27JLmZpݻTx.uǪ͉@uwݢ\[feC7EGW,;_(Hl,$b)W"C(S۝ XǕoX.P^ <"ՓAbi{.Y;9$dnٲ wT?#sXs~j엂"CDkG0Dc.FMA6b`\Ǻ!d-a)v $GDvl}.̦R e F8u9 FqjH6Q׿1wwyRw%hWBI^q,/#>tbwY9+,zXmO_]Cf$f툖;9^xOu 8yx?oeū-Z rx_\j["A/rH2"HMA+_>nV\1قxs1:#_]=Glb͏~H7uHם12g\p3E+e\g.*2esd ǂ{[wz@$aꅷeiōZi4)3Zȍ _{uJ2]hȸ#rRۚ4fZCF2B V z*b/Jh*p G8vG,7L|L.xzY4nk\4z8|N7B:O89biRsFUM<sͥ[LAqJS)JlW4X_/lmZ@3ʴh@Rmq57O"XF,,0 f'R<\Ʉ3f=udT/"~u1J4Jא(*%$(]QX枤fiSDS`+ o:dՆciS!μzӉ࿐>yZ6m@|͂vÅ fRz +,uSgɯ~/lkOfX K AkXLr$cX\cȠR*Ÿ?x\72M]?l}Q, YE1[@e/}YDj{[s _ H Ӕ3nVIJ{m.; N8m8+QV Ƒ,-+rr1dRԀΡ2mf:yRL"6K0QݚT_֩&x 7l 7 [tuKaG_۵J|V&F+j( k6c7zXcPvW~ %Ok]DkF.g~X+0ZU}w07oHzL ]b9D딩 #W$F_+qy~ KUB/QRoA(׺q!`JFCd*i"3P:NwHΛ9?SU%ѣK ayZeA:+Whk.N~xz ¶L N2"o٤2&Lb4$~LGL:s[?H+ ry*UU.d7)c^jXf< \+%M.L&rB  P¤N'[[ףm_vO{Ϗ!ɕoboo/IG,\qCU!S!@Y^(?5$R !]("(\E/:L<` sJ`@y@JAn=g[XX >ؚu)̂;7)RqVr,U߾2!PȵnYpT_2[n*SiH+#ب$4-ƦrӴr71i¯fP ЈblC-h72+ƶқTNN$wc{Pn`H3S'Yiv]T_֩R@r $" %_-2 ;\Y„oFrhm=._)V5t2Uu%;Y'/`o^uw?T? J);>R#es͉5ǵDx1=ROaP&]{.l3隱#I׵1ҌU׾,ĥzvOHm W9R'B\Ww]HӸP\]6`d =Id k RϡZˊ*QߺAFVr0U VѴ V[`adT \D#T 0F1:,<754cI)'. 4PRD!%90)rȬJL.麔Rd HvQ}YиxmȪṐR OJwS}Y֓ssR()*d[QEI;N64.[&a#}- OB|^ y*Ղ_h.)lFjV""^rZ.SAF'4.E ~}-0Çj" c.-v0GDQLx.Qܨua2Ϙ@3r#2E\B9#ljhk,`S/fg]08Lqv mu B^` ZUY.I=mÙ}jNָpm]ܦj @w~gE4E2'kpt4cvo;v'|Op]ـ`o#XF,CJP {|EA=J[d2OL$[y% :s,fM ]e+:% S6df3^dX(sR6'4r e.UijSO 0%p1EBA$IXdB" i"N2 W-49\k:9t+SrFU@ Ǫp=&LE[x:-:Yu7/*{=hwI;ɏCHʓٛhʅ`tn4G5$|xT!Sh5L3&Aq64\)@aʅNW"d%GRT!%o1)*yv>z{7kE`)T8[|EVcsnŽHix.14N@JwS}YPm [J-I4Mej句QJZJRQ^? )*NJK{F)=k)E' iqR:ՈLQJOJ7"4<7Rckm$7EK-R,i&/'0nX[6$yvvCd%V_Ԛ.;vw*XUs<ɧfdI:&Sӄiv27\&d% +FWkBLo:#/ l? cy=%iU`8/bT6ǗdԧesSZa։ f4jf(ex.2˩!4K<1<qP@GIHJ!T˧nƻj(%4+׏%/.Q]ѷ!$ O]읧Mcs i 1it(̱m\׏cNbƭBG3pRaLXR#8R4'LBgn4ha8D$~'8*wF"QJ~pI 0R圙j+"" 4o阶ێg+ɴ$.Z6F(%I6ɓqdzL"E y 5иŌP Aٸ}~=kQW(b}&NH"#ZP}8Q/%ח]R3 {.o?x]h^Aea IQ ƺ^Vޑ:H\m;6ԊHx[\pAPhy[z@v .w;mozr;4.[nG=v1~EPJyGtTi[ EC=8ɕ@䥼6z]R&y9Rf`2޿l d/7϶Hq(Rǥ #0XLR1<% 5/J3]לŁZ:Qu ֕v U!`Bwq]f+;x ̢$׺U`* ;=# }^ .'6C5!}߀>bO[_eO[kt\#њbۀ̰26nnc]˲]s(ŶR1u 89ۅ]/7>\tJI.F#)UgxO}?9z*%%;\=9sOwΓcդGO&]$t|-~jix&In6}Ĕ%}'`WMNYZuL0IҍgO|ţY.L<&~}2 J{QgUnlWEIPR?ND Kw1kts6)4E]赉!Nu*:Z5:;%kN44^xӬ6aEpC +Zfϓ HDQ .֢GPYSH-j{yVƞ ؂ b'@):R[XmkI яM7=,8F5$ TKWeM֓W68>ݛKd6'y%M:sڐ}*zGA-`g6/Ji.Ӆ !=Bh][$iFP7C"UHҐ&ԝбmi@vL/;Eiw_k(=iMG1?H}Q~%鍁&h1 Ecuga^ ܆(Gb84QR}BZt 0šf%/qd?-Ҫ SJ+h8'm&DD8xj%bק7.Q)TdyoTiKh9pq$^iT+UUwA~%kNnlҴF' ?'#GCB@W^J c_F1L &3c,KLB;LN ^AdV96&UPe ;(0Sy$d MHrfiw*2uPXd]؁ې;IOM ^7>ڙ~]v_>=Ȟ+H30w(}ˋ߱! pMQ׿a7^ax?u_c:=\!-!Ɠ|jFy޽{ L{$L} ^i|ͭ2 6}c%-OSo-0k hU">>3ރXtM-ftCU0GMx`LH#eU?jKXI⍦K3ELBeɎo6,(f Ӊ BX"υ8@jyQ܀ -f_lhtktQ"t2& Ϩԍ%qh4eVE$ 0@!:y+Hza8!Ӕl)춐9G 9CؐCь װ}!q!g¡uBAěAo&C\ȹx5Rmy_}XLdU? xQNvGLƍf:2͸΀Ƃ \0KtʿϺ\&v -(ge֫<ՆsӌA$0,]R?%:7~/ΕՀE bt/LKecE-hYT%F`eL;,U AHpĐ3JJ-#xiq"-j`PR35e/K^O7u#[P!9Md[A`8S_nm P͹ˆ wUy+Owů,dR<sD'$o3B0c-O{cjLGc;Z~z0żazyՏ(CJj5s~~W<}?Q4L Yr(1Oc EΟqwzwtd)\%I@it)kЯoMn}R_3d j\`8.Pn0}_ڛMi*&J W>Z."\Eگ"z1M8GRGU*4,|jJoUg'G 9 # ]AmEH3 `+KI3c%ߘs_(H+Q̅8r̅VZ^!4rtvfZX)q*E_HX&/ĠdP(z1{Wu={߿m۠A?o|C:?&[N㇢%'AoB +Zn r^N1R`Jf^KFmBTnChI~@kyּ eCru#LTu QLlT?JaH;iW\XNqQoJr.22UA^vIi8YL@l-S}ϸd^qYPЯ}D_]%Bʶ@eYRww{{9R y}^xzy^TduD>9OO{Zv2g:r0TrF&krIL.l 3"K6ٴsjgmZO5D~GSPn Sm+S['_?ήZEI@mr*Nמ}Ƶ??i _wnloʈJj*k&gK"p/˟֫`_>J;<%OSLeeX a*>k K^;T›)5˰/[5mY!T8ҥ9qG }B57X q1h\4wqfC^A tY$&:@t++i*Ysib-VܻF RA̻VhY<-;aw(.< ]?0zבқ!e J']u܈CGssWhNEb56eTK2{u'S%zGVcYwAzG_1j,NmVNCəH\*c܈48jr; м_q[Eawت6$gΦ -jk$ʣd\0!\(,r_B'@+=f)d"0wRgcf3 ]mo9+?Z2_C.,If `7DYҨ%'Y#-%K6[ݒ팂 Qŧbɪ"r/Sm.qfcqXE8nzdB4h-9e4ւa46FsEF!+~0"RP0f(5B抠)FIy"8h"%&%I}vS*Aݘe  tl\YeP % Xe8wHA PI &&p1B.",Oi.ĪNEPFLbz, Ӡ E䌬<˘ON1~=rڊ\x.ėCfIlX z6hd\jJе.9t 잭)~.Tqn.z'Ǜ+*p 2oy8m~.c0 Paq4hveiG1!:`MQ7 M܇~NA{=#]:wFs :( \T0KIB*d@7D;+)KOGŸ } uNԇ -S k1gӚrw%oc޺da=pqK?W5tRCʽWU$lu\QT_|Uܷh}npQ^ϡ)r / /ʛh)K JQeaisW{:K0mcAZ ڛ@­Ta8R&iO۝UAUJ~eO U [O?* OSI &)"zSEBZ%{'@HpÊ3vBs\~rAl@jB{XQ!d T'V$G} `TV[#4Ƌ$Biǒy-nB |[n/qȾՈJz` #M A/*uIkt ( ZA@&l^AFA~iK^(܎\(>!𔋻s-p|]:fZ]m|&71`ϕ:P{"u7kŻh>ȃץ]1Q5 4t?+TI nmxmҧUU!(lߧat"(%Nwkhݭk}*2e]ɘԲ[jzrvFy j$Q5>'C+a0",n ˩qM8<׫OuSER&CAq"ac Lmte۟T1nS|=X@g0 a Hx{8@; _μВʊ皨|qض\-ݡF!yiSQ /?ǓCR<FM*$)aJ8#e3P qC% HفC YōҁJJ׍V#].0ֻn!aD)M[%uj$ᩖ ,E3G,Qe/2^jn( {9I3Nyz;qTdB|V|YM0b # 7Y w/wZ9$ވ) dJFl`K2?ڈ)'ShBG-(Bbn?WsK˥c*zۂ7BW&Plp$4PR'Qђ]~2p :Wfu2ԄnP>Λ`ڱ!:lp5; 1&A!NgttY@i2C0݊u;:} ;ϼ4LԂ%LG[v$BZ"}gYao#Jb÷C1E=`|JbPziƑ<p0뫳t{7d41w1Z:D1'G2LXQB@)ƶVyC( 4u;ӭ|)>)J8#1t9 Dx/Gɪ?Fzh &!9`2jsdž=WzFGsd +GۣH' Z{ˊr*J:PN q$9G{axB>Qx_XZ dm*Ԝqrf_p."qmΣ6mф<_$ߞ_KWH;?#B8ɖQ uɨ8X:?ۙLܖdW-0MtI-eJZigUr CPztu=/Ll8ʦ ^>N.Gջor,_Y^}/Pz~P¯7rz|/{'AWrbN^qAnަ!BS `޲||76bxW brׁyPQk }6|YRe0_!b,@058X M8K Oy4@ Cyř-O a.) q]:N20ˈg'٧x WLߝY0injoYKZ1<*jZ03LQ׊HD y}@>쌼?ϊpsbwZ3CYaqkJ\t9WH5MSs-S/vƥ*d$.{i(JnP*&ZDلJx kau`JTNјa}??$R->bAFoqx'KBR)`xPRā}6qA}'J Y sg lZU17}/3"˯O.(%D<0+ "򑅵Ԉ m&\S7 аMd[+O#,.tBP,+Tൈbp iF-97験fbtvdm2Q)oiEFry% WL%jzلpIW[Vh_)MNjdڀhf{%:_RyBl? wIߟ W_7m!CvTBhG42ИemЬHT#}D5د;ӑG;>d̑7F%!O6 XVSV6±XC+TҮ˷$b~hF7$`8ܛn }C|j,KH'QE0sC 5\)B&-,7{Py9IQK mfkxh,{ݓ&*螼>w6rzvRϾ ffj]tXXx@%$ ~N&]p&8k.@ԕ,[o,[č߲J'[פqR<Ȧ.R-Eg~⢱I:C6"n5mteſ^vK'o9k(Jn31L'ьw4Ѡ ۳Ɣ1käʬ542khM95tOYCUd ^W$ծS`3R %RN $ƥG[:!V+ -| fquA\2a0=xAplť|r t艳ʭοx*zNk8[fbkGi5Ca.ZR^7 {73u5X]#7M) ϟ+0ށgu)wL,0R|r%ʘj6oԝz[42ba0>Tn=0Qz ",-Rg epgmJ: \PX-*5e8HնT:"Lq&ܻ~~ZkHrqcA*S'J\%KZ.'&M,ƖTT$M` UYv.3Dql4Ctwk &r#7NRj IJαup")*  /Xxb2nH"Q$]N@1uXf&¼(u!CKJyʍ2iXD<7P(.Z}I .jAy^IY ,[ pcao sopDY*0WJj&:A0h^$u PS-""஦+ʶ ȩj/ Wk (OM'\5 Hɸx9L`.Q\KP >k Xt JD$5EE*, 1qsFk6`bzIVZ p8?K Cj=3ر?JQ_GHbX_ ӦٻM5Ż1h׭?!uf4x*Gaa["Rd?u 2xCNZU5{ FXMEtuﻦRԗ| ozM=“xkb-K?'7|fkz`RhC=R&0{Z .rZÞ 1<=-5)PSC(^ aE-/f;g9CѲj{>~A{UtoO>}r<3&̴qu}nvu _!f4's?@F dm2!In'⏝n'zo'Ϋ[=W|[ָ CTF`sǻڳbeb:ݖ)C8J==>KY*qhVh%w]^Cg˻amK;.Fp/-üT۲[QfQIE?cH#p$}QC]mRE9D\p˧M8~ϜE❓Z6IEb,9 4I`! wë.ӳjcLuI49"N&Qۉ5Zjmk fdߟ`C9LwB|{Ï@5/\mhi O6#y W8 LKGL[cO%rd\8aR/k`7D; hɲ.Nas!+8t8WG[`EAm$ϯk˕chKu[g nj7TWVlr$s/froW{|7OwC5N=SrȬ"=rHՙ2qt^cL)oW@O6Ee[F#YuByeɵ#dJK϶ F.E4ˍ*w;@VJq]id-,c fױ9,EnY31$9b[p"ukfk_s-`O~PmOͮS[:4BոC5JZPX[fC' 0Ob}zRaxaO& V(R p` 8ؑϊvŷVn7w5o)N)(ŽVAv z0{"LTV;@Dɉx2rdĶp2=! ;z3Ϡڸ˾$4# )̮TB i SPhqm3ǥY*q=I"N9F>TI&2dwIo XXT z?2$*)*wG+|RɿiIx*>so&pyRMAtW>iVj]AksE_QLGֹۋ@!n# ۄ8u1}@]WXVڮ"Us]Åj^g\:ph!clWF ^gf],=a KqF/'D0a/3Nm +v% 8ND>cT߃`xDK7Z֧`|ۇ4ؓj8jm2Zo`W먅 F] +@4&h)vǮ0" CǶ+Fwb(D\*85pVX-f+_ v EX=\b0׸. z8!cRl8YznI Fu|% gG/!2ӴCba<4Dɇ S;k3;D0L,zeq +뼏@*Ui ISkax-2VmҋTvVB ^K/.B `#,TaTK˶RaV ՜F%4 Қjp-\l+wyR kpT^j ,\/ufJͬ3&Oˬ O&e9ԕTfYR9mZE}\8l9 oٚں;hjJ[uV[SjhìTUg|$JYiMK/JvPCG*6T{]` ʜTCt˳R ,wQM0Ǘk# حV1v=+)piFZV׬h( V"u`B[$P#y$% \y`Usy\{GUU0$mXwf&O|$g-Vb]P`&UGtDZ?w h)8-7H\ !&Lp]ϓ?gj'i~W>)|~޿h?FuwT6r. Y:!4봒EYVe‚,\.*t+ TviX_UY`spur4j%fU8&>QJf%-.9PJ/{~*Dfѐe06H̛"XI!u5.p~{y_&kTLD4v^P߫}R|6f ģv6z d۰ W2@ aـ~ a{@5։wrp!@0׸= k臁h~V7~.>@^`3HgC>vSfR_4wQv J gSKARƈـ~@+HL4sLAbE^E%.Bްd`>Gz,5C9qoTWYk E%*ժ}WzKy|}rJ''u7#+ i׋*|@q= p ~Ts+Kyxcߏ3zZ3RUWOv$Y$E~$1P9=tڙ! =P(~ӱcP-[#,sa_A>dwk^rN%]{Ζ3p1zmJNۼҒw|b~ `\'ݳ^avF!F͐iVLsnj-K=evead.rHBC2&pN_edh?`Z Z{A-(MR@;RU*RҞZ%,$ {Juh5ϗ?Va2q4޺9@.dR0.Qpb%f350(ˣ]*!80+sya-ms$2Q8r0I:#Rgoc <ܣ 1e!PAMcU0[YJkJ1z, )Afer^sUcaI<|C -Ws_[\3Gb/j&bnհv.6_[M7)7????os}Ru]Wd@rWh*&\\s*a5ZEn5˚_ߜM~r%i񠅬.=A7Wԗ Au qVH[ghs6PV%j/47SVʖtǰQIʉOT{ĉcyrIfeGD`2sq,%yա8}^Z6E"M $mE dByFRkBUF,л99E-$j.h<@j#3&4J04% 5Ж 8ϤOCyy2 Eh;/IE&Ngy3-eT;z4Pt2؍I!T @Y麦@䂎+t)sSB)gV:Sd-H {;8~ZpaADECTRM~ ]+ ~f`NJ~/|Wljte ibpfIyTެ.?b H5IX>߄ıXܭbl$Gglny^i8`, :URײ~ZXI`Gpȍ49Df=g Ji /DK.gɻ· t&6i}?^/9BI\3: ww=T*aT݋-cϦІmhnF(Žx- Krcf7JC7a'e]wOr[ҴwD:Ymz|oጹ}L}6KW.'I)7gv/"^e%|&f*JDीdKRI޺91|M7'(oT.27G嶢r>^}\/Cw#F,cҴXX Gx#VφUamV"G*?E6+G|d9y]>vRm5}x_W?ԡvŏwh%=ן]vD)F{eR0H~M>s= & Ad BfdVzܹ/ |hQ$-{<Fy`ɢOZ= *۞2IMZ#k%*N.QIڰ}1@iqhzTL%Q 4 6hϟٴjfo2w7w= l~]ly\wZ ܢ˫Eg T\_ya:ar;kIuKh݆АuJ1ߪ7qbXY!O3kQ,hKA,q2(z6l_;fCTR-P*VAJgVr0b  4d+@ ?4p4vXs1hД| H;ɵT! o(ܔeϫQ*b:WL nRh Qh ϬuAR2Q~c]mΈe(#Gh)12d ˁӃfu=I6#?₾-jO=uy8h!a Q2ƃC`ӣKq7A#$ 18Xs9Tx~0& *lM; o#?-Sh#l!H)=TdHi. 3pvq TQd'QĞ$Hi(b (MIn֟jCfP,ڈ>Z3v&}j[UG svА):%xݷn6NGn<QwX#2$NCZm y*S'q޷nFSn<QwXcF 8uKh݆А "d^-5?y5+_ 8 /ȌR ֭ o59eQ6U:MvWE*81 IX*D N sX@I$?XlvYjy5-)wN4[{b?a; 0r9P^ο5JSqv ґn&".BI?׶aS|}Xh/>LHX>]~inTb1\],] lP8P/aHZ d^I"i!Wd%ɰƫz5^tC(/dΔJP`#(K!q:;eF a [RvHN(͢Op9(Yk A5aCol箾pG-땟ƾWSzѻ*Ti keJYZ ړ$XSA՞V Zg  㔓l"b80.EɄ+u)k@FcK^khx]+QqIMӯ"[\̈/jN).Lpvv 2BH1A"8IQocAxؿm!VZY6@= 3-fB+/?؇=ט99@Ot&[#;\DZ@r *}K _ eu 8NSFeN7_dk('lZ؉zx3`0j{jE27Zq r*0*'쬒g<.\֔ky܇4viE3({h˝HwjrIGߣcd&5#3zndTMVݛpK:?/gor3F3yDSY~1YדO4iKrnbwNiqU}{dЭjw]Ehw38?x߽a0wnKCtFv>yCOnXudތ!mY@YFuU>ɠ[&iT_. \b 1d#gښۺ_aevPt*?%}TRˤTZKBQqI"E )/c8_7F>D_)ݵ*,<"yC[hN!Y"F3cϒMZ EuD.~_1#ϓ5xGFZP:07iIF7ރF+Th`Ƅv6ʢwHeѪXi}Ezx/\{:,l Mg`YFQO,xi,d$ }Hj/6`V+~"xD'KnqEJH@{di(,UT24vF %Z"ZAgc`aㆰ*摟jEiY4"4OU<*cA0^hcQr+c0#(I$y2lWv[`SO;ùGP"LA3$[XΊ`x}얭KtA,N\u%H.*!Dm \xAD LNSo,:E, О; AՍ+ hkcKR踦X6x18H /Eja~n CO$ XR%Dx=sUN0VHdܼ#]L.N0q,Fi8\2 Kwۗv4Ŏq/>-Iݨ7>Ӑ'vw8dEvbȵ F!BT8iQ7wo\Rw@tUJ2F?[ E ' yrajUv2,L)zxWz~zqZꍎIpiw}3Kdw2ӤQ|]q帝IJlrw2vwE&|xoW(_h}IPMf{;ŷ|t^,!E8% p^wJlKp!cK)OjέNH^|`c Ip\M.rǟkG$"-lwӢ:/lBmumэޯlT9d=c O|H$B gCBxKÞ:}kɝ#z95tc/uӽ=<Īmp *hH䋄«9(饉<AS'-VO PprFs grFfP!nvsJLB,B2QdUzWNqAd=r04r` -0Fl3T B:nFԫLm\NMEn5T[/.Dmek1DG:U=>WT),޹NAU0Ǡ`syS/ݙ#HH)+w>%F"\\\^ͼQn~@*Wh/5FY7W@"uhQyH}1JO*42*JbGIdRWU0u ՎwT3i.1D>$3Mn>2_bDCқg㉩P\iEEs.}KtTRBzʩcҩi/f})u ߒT<I 8H&Eߺ[q}'Aڗmq\[M"f^lOFxq:81Rޤ}Q9φIP s r#o_Mj5e$7ʆyInݟJ-Z^ejyTnv}W9ʹ,r#$Rc$68@Ȕpy)- k{-77w]]lٯ7EI}1~[R$nV_~6 *Z_S'?8 IQ3Ty˧L39>ȵyҗv&S.`%n=[bNK]%o+Rx Bq`TBHEH8Nyb0MPJ9QV Dk1'uZz<^VP xMʃ^;D5$=R(u #(rqjExjfI)D)t)̥I.EfiOT1qI*5QjK\av#\΄  y5{r!KShW!`鳟~v!ZlKRfd~ P)#r 3bH G+%,CiZb+m u[MPZ=cc? )5 },Cd7 Ct6َB4 "B׽,6wH{'ƨe Q44I Is.A#y^#fYk{rjٳ/PK2+Bc@$hn,yʀP^`CA`v#X@̩$߮t~^3z ; }*pPڗc-ZK/CE"bqS˾4eqٗҦ 2hNZPo?2diQ<34e)K%A2huѐxi9+k+ŸvXw߽Zkт`t;ka`QHHs2FޮlU &>OfgY>8®J;$ "FVWکX)RIwQ_{{q]\/3%#\.߽^y` X37ƅYگ J_> rn wr)XǠt}SoJpAsU$-r|b6T7wsy, H>7U %uUF_k,xT]7u7KĴ_Ԉz~zGq_P(@'g{^HgR2j(ɜ-U3k3絡ׄa qy9 Rrԑr]HrNPSG'>=֚8cQd%tY(AGx)Sۨ^ݾS`BMkY֐pK$3cb._”C(΀6QbLy:P4L pYa Mp^HXl#hcDg' (l3;-ZKzu$ֻQ>*ȿ( VqM X ŴdXN{fh%$cRP}HH]ۋ8)B)?[/މ${wvV (G $gSbL][XV{ .qjpw+Aiٳp&[fsUҋw_wHz-LlFLUBf:$||yk܇4WބۻOS6sU7LwkzLJ\%/Iu-LJ3hޫ}wqlC^K\,> -z+d/ rH2d;2Ly#=FyKb 7IK\! YfdL)urO;u77&+L'r ̐kU%ylVv%޵#be1%@v,X veYr$9'`[ܶ%-Z9;UUŪbϹJ% ./sfAe  J'u(h?ܖ<ے\DtzѺMs+¸v*;T]tV(挍7 9KOtX~O=K!uP!Of7a)߇#сx٩]f{wf!Z^eE8Ԯ|/L"+j(~6zߩ 9?e}A'!opp4|`*J9Lٽ<-<Sxê=G[0jik-p)BǍJQxG_!!B$@1l 6c"i$:Gǒo{L@5-!!mo/琴S7Gv(9C+[.wNa pC/r9rM{U\?c0|/ԃRqjs0 %l8&R{u -^LΊ[陎,t^~k;J-7Mu_ZhAmʟN?_zA?$t\>IZ؉kBddy?#ViL!m:SgI9=T~dXSgf =>*跙BEzײv-*{Y6O(߫Gx;6{ss_0Xu(U~DT8tRpUM +C>W}~0PҋLʟ Tm}H']Pzt{WM\1nX}p ';녂N J/W78~ֆJϛcn^DtInZ8`ꉉL16j g;Ԑ8VzDEswRb{nEQڏ=FKU{ɼ|.חšӣiki\9WT2zs/ޛ MѡJʿuyWi&\ 0Oqny,nK0C\y/faS9} ӳUڿǚy𰈫i+/C<5mbSX5e.l?dwTHF_|VnB߹/xD'Iڭ%S}b@r~E.1և?A&O2'ɓIr@8"IF)KN9OpqS<e`3#qZt#-k{|Ԟ >$^*:%TՎzӣ^SF=Žt*Wrլ;9eȕPIa2kY%#w9ƌ|~8rQY|:f-`~8tS`sq!HCt9h9 SB \g^ِt k!RE'XxQJOSZúDjw`$RB}gi-FR<2Td`4E&6oC1?-!$@!;)'ÒLSI! EN I9J2 |BZk0v|!6ͳQjJcp'rUMCbZע`-){f!og7d3Pj:ߓ%;$f͒gZ{%“hw^2arNu#BR.)ttigȂF'A B7_ ;aA4{= s]~@:\c"aʍD ]|0y  I'JfV")pqy2-Dƚ]Z< $xV~mĎzc@;#2Iā4߫1^F #y'S ۰Pj2Ƃr v!(~;q] zdH2Xf2DhCbCk_*ăd9@ <- @ Vf$B#E$,Ay0R!C.)v~XoI2i6<&\#eDK <Jپ] LNq.?~#'4Pry@GcރR(G@!NBY[O$찊i(5sr'xL6FOŒZή1|fK` 5$^m&7 1b&,Ҟ "8RekP 6ti7A0Hg+ivGj b.-<^?(G/թ.SD֨DX(6QI"$7kZQ o6u_uҥ7鳞(1Vbtx/B_uBnd@j <,(0pH͏_I K7Eg]"uqa RSl~B%D.I9ft*&:WҘIqlܨUk0Uۢ (܁:FxBƻ_?O]]\[>preZ<G\^̣_~,1=rFE~i~>A==%kyL*!q` n IED Yəzao\ciy7kÒ }=_|Gf8'~\$a'ŻrJQzXh#8(fYlxgXT,"Dly)IltB0lxS8d-52ڜs n쓎3f5.1 vFkx{RLp/QvOqv|7jhɴ\^jLGerrH14pV6OߪƥBH\'E1tWkԀF<CA*ĨBZ\ +CϬbB )T2.IÞ{xv| H*V(,m?<x$ 'T un l O|LQ̂1ҁe/Jɚf@P *x|$"R2kJOi"Q1%LYs+ErȆbQbfn?}H̘I!rG]B#4 T.MBR^WSF*2)X"׸M֖ 7-<8J~(Fnf"Ʀ|Z%|/ІwE{XNB2lhÙL56FM, LMFsٻ6$WN'рfma11DK"ii^߈"%d*nj mQd28 .*C3G;/M2oYb24qp\ŒYCP? fY *QRWV8ճ!1nb.ZW<|l\4P-)1&ZeS>;B-GpK2گj2,-)T;̳w$0扐<3Pu%<}qZfeߵ }֋ܟeuJ h"s|Lb##v$~_M'[y;R帱Gkv}p{CcBؾ%MS˂ys/Q۪?#R(u\lGgڣ%plw$&GG%0xPw#5?Mh eqh֡ 5R\]ȗ>ufq_̪W6eUD}*G;-?el"_r #*JQBYQ /x$k62Ok׳cLE1=}Zǫ5km=TIEbDdɖ8Sa[FV~Q%tk}cT;XW<)91:y"YIrª5'֩ 7dLVc]D6uu 7S?8؀a 7֒xP|A{@RF/WjOm fɝrLjb70M NH1J:iKZ|%S(P@?Ǘ8j><6hmmw5= |ÙW?8.׌7<-r7`oX`)΋ӇBVVuٔ bS{ƾ Z(U/u)'c$9ia.Ax_}dkϧqb,@"SmceiWDvā"\ު+9lOʦ(&.~XW,;XċqƗP$`<)s]8ʀP*eT$kYYn@mp1 [kpT% tȠ rH P g,B_PFv+2Z0)V o:.㥴DK"< =8c6/qdr̤Y{ e@(Q)pmᛎS`amgZp5&@C!rL_𙕖on$U3e gI',]ML}a=yӸOH#WZ/38ˆ1!I{3(i%JÏ lr `)NV"vSٸԯf sv-coB4.1Xb/ s*Ex=x; lN}#|R`AqEIb֩2:iJƱ؁́ݳX{ 39֘1VgDcl`˘L5T'XɑE#}?eh;wF`$|~_W!bz%%I0h;|)@,_lc}͍|+H1{x8 d}7ja/؆?S\QW9JL8(h~ jBETCocs;8ǂW)ϫ!b;~[`=/`]7(4{8OhQ&Cp\FFT\S`w;טF L#ťIGGM; Ӡ=XTq+|ضǻN|xY 3ԂIٟW$+sKoWp!Uzc'rX$K%DepN7HcbN4CR_@Y쳠竡cO s7ڴs!7c["CȄ; ͽmק_P]s?kݭ5[a)v}x\]JRʆTrOg[Ͼj@ J\$^QmQTiªzG| %H}RN(a 8b z%FH}9V)e .BY %D+ sͤ>,Χ )$9OItܚDh xNՊGڀD;a|ߐ)!W28kWeիA=]c~qnOs:C_.$s)ͅܜnOVօT-U;jX@+JGR<4 Ʈ$ED)һMCz)ZYsX蚱:X囷טDze`6?qώ@QR@ytPI ӛRA,M-$ +[X܁L`:_عAh.sr(.5lk(W?g2^ ݅odt Uv8_YF[;lp מi7-3r]\\qު8_GcpnTѯhyo߼˯aS&J .A]O'79.|zyehw6\^aKV159?eF]m z8b]daN`Θ0Z-W&Vr&_)j6uﮂ™*Oٍl~ydp+;FTMQ_2-<x;K LNys(qVzelX 4+;na1zP˺9]=zQ8˘ֈL52j)ChF(,82JT򖇇@ޑ|(_FN By5(xUιl4&e*Da,Xp&tTL`@elrVG9*,[5QܷiRLYXC_v <8€)?}u\5z9o~_ B9lz3o}y2(yXsͲidz3C-HӃ!Վ+:\zq{e=>d9gu_k +bc9ZlzR8jr= //}2h^w^tMMk/4|jb~O _?y(u~: C96|ή$`q>lӻ(ka:Үn\vw$3m Fh,e%_Gsx b :d,13>r_h9XE^H&v#|`߅DUB:;@ϓb'M$o}$U?F8ժ>(`(VFao ӿ{58/.wذeRpq k >SED ti & ?DGV٬?uGzh$eJ5_ӨF_^f?`:7ƃ~ [yWbx]M`q+ iJ*<8_^F~uea(Cg߯Cu(ݤ[8QH]>޵ᭁfߥ9JKRg(oP2SWUxք =~K֐"Y`% ȗdsgv粤yumV {%Q p1Bo!I%C+V)[*V d@p%+Z=S;1FW_vΈĐ =t9 AKX+]BG5n.41e'.%yK} 4bTAQǗApƅE+fI”tY]Qhx?]PǚDN_Z.=*# UR§rp+̕B}`}iej,wa aiVH@X"0ޓ.: uXٯ?dXJWUEE]makמۊʳ*'d ;@mvKW0\f:jXj1e3OjT dJ0c Q`&u! F>{Sq$.Tʗ JQG_a^*mX!QrQF8i+ů֊HIy&Āk4 fa+B}:]B짻4 uNS$BƒIpNҥ,~8Q `KmlL$"Аn"TK K&q`peuhl΄t#O} q&$2gB eBd8 KRf6lCH==oM (~w 994֍\BHEqjt}}b+l2=w{ZDsU{o^#`'eV eŹHg}z;/Bt) =å#)G3hNou6-]m<ĩ||!58^|I lҸpf-,@*gucuR H[d vurՓcռvBmtŅTT(DyA6`*oLWIoIPK>6dbFSR:VƐJѢ۳?<>E]DRY]Q) -E0 R.g3Ji;G,~JvFU Ի@0e jmPD5d6-nQd8o"<_3eIE~k֨;nԤ,'-\AMyd}eZXQDa.JvbcMjxSb=^ 8G:%2cI=W!Z[L_; ]}៦֔Մ>D ?O?=a׌Ѹ~Ŀ4v>v4{//#:'3t{U{qF@szR0+`ŭOP6H0Ɵgy7[3- HTdS{b2r 5~w>*lomüIn 㴿1{<8uq)?oѨS`Xڌ?oE-Hq,NG+"R;#.i6iM#ols|}/mXF6Gx*4V$3=r [J(\( x_DRyrt3-F\I#!\D1~:yIXpw֕9m!,EloڭhvCB޹n/SFQ;KlOǧkc{oF9ATI9h$Y]dي؊5vWq~^׼hIlnjZ'bv^Y^kN,5ܼ)=嵶RZX嵶 *Z;赲ҽ 6I 1zr4F,;oh'Z-W'ZB4Jl^7!1Du ,90b]Y,fJ+^Vuߔ*k-9}-9U~M|+bƁ`Nۛz{`RԖ$v^ҳ\ξob~yf 5hQ۫;gNaKJ`&%pRIo5}caB'̾'StƮ rL_ c꯫e{M-h@4=N80bKԥdU7*|(J.}N( -sRfHÕ-n"i-ƿdI)!tԟ6,-Th)FI\j1/ӉRK  V9]UE ]1 b;VVӰ9N6%+SKU+F J65d'QbO9ޘ̖sO! wUs^8wĉ:Ƞ潯- i컰謕p@1=s଍^*謅2pu|v>Z!9d %VJQsNBiZoI%!PyEiw4/5_ Gh xPT9rR8e7d#}ND}_Y{ܾ3)@bp(e!D+`Z& CWA0mgN(dƹ*BCu͙ќ.~¨䩧D8|JUW""En˭ vSyiڔo `fVw(R` ;ց +KA4W`J67TeJwnXOC :ӿMPZ)͋NE~r :/A+hsA.3@\׭ӵ$/KYr 4fiEjA++JaEQH.4đ]V+PuUU2ZGW%S ;dA%/3u3e\Z0.L*^Y<5$"zKϣ!IjruQ\gYӇ?dָƑqa?tANķ|X~t]`f4lt#n#e_kRaqFEhoEr-%Ms<ɋ#[VK~x0H].VzT~حu+zY^4?nwr}ifKa`a{986uFu'i+u 랞Ҟm&^~oWfoX呐D@n !ICR2R(}V1.QIBTuѝhJ P UkXGcTsfVSœ @VQUǦJY]Y "JT{Z9%ρJ  MϓW \ Aŧ =44i@юF^=1*HDTni ia4hmt<&h/Bw6ӐV&ϻLoAjz?jje45\s$P?s@]]@swܳ)|F3]џj8k:XޭAI,.PYlI'b)2LHdbC"%NOy6wͱvL9\}2nVKқ?o66c֭lnn?- Iryë?m}Uq+7@U> x[ܮU#5q7C~n|{TmZܞJ!\EtU14dd CyF˃}FuP跬ɬ[~upuBCfX7d"[.bT')mƐ"#UN))Por͏-DBS>H(цjΈ~>~2rkn(r((|]nn ^+;(sYe~qjPjD{Bc03&:.G$UAYBdNY(:XYBi e%pFL8`@JE0╧URLm%Pd} =Z.klۛ*o]_~G,laů_6?g* w”Kd-6nusXdAF(gfhY|#`*BV[ұBj BB$pwU٪yQ@M7_dOFZK#ȦlRH?F..M<:N&sW\&8Ӻu7R?[u $ J#e_2F":\-(@] P})0Jy 1++e8pݒH)+~i% ZV[3Xahᤵ㶶 FQ"ʊƞ*+NeA;R_ήH&%ج5:h~=a)&sҽ kNסןz۫w;qq׹uqlޖA:$*NCuMEC$s`I4Q} ;K)1g.kq^vIj3W?h)`Q 8'2*) 7J`z]xQZrZf 7bppC^ؖҥQ_@TP-ظ]"5j9EFƌ/p!E{hiRDƅ߸6{UXFVj[J !\q2?eT#WTP:!`]B0F6~΄eE8h]Rh/֕&P }MEaSʞ`LX%SR K!ѢsVʢC+3\jvrVk~ #68o;*Bӗ3rۚoVD90\ `if0v Oa`oBTjUjY%J%ώwrD>axHYתd T ˙Yݖ5higxw x<<<`kң4F}I:pq2£ZS8FuuyaXӫ ^3Cpr-n|y\Z!SsZ7hu]P%T Zx]r}R'ƴD*0ڤoGP'c$tOW'zLm D {"yGC&*Twx * w9?ކBZ,nZ>ԹXTnqO N-|]Pwmo . -5_-VRcQ)j`|%uF׋zkDpEZl ]-v₈tPov ^]kL)kgG#y?7cbr,<ȍB$aI}Kk`b%˲.ֆfTc(ܤ %,\B%5[C@xj*iVaDIe8UR0Q0%Y|ԡxruA|rk̹)- B9?;Iv/4@A>\-YUZZ|Cբ7~SNīݤ7I#^c/w@ǟ\aS!T P2v0+jI5rMRb0K=2,`dd̓QXWݾ/>(`wךh{秛٤B!TPvPB@ '-kM9 ,20 )NypP,\d6מ~F i MR-Wޛ_\/HP>>6Ӗ&^~]4]迚ua9\׌*~z/MIQTyGP&Bi2FR8?ZD¬:Y{i,YXi 8JG,t`*P14ӄP˥ YXQp2m'4nOG.Hȏ Ώzi(NvQ!aQ㑛fqwOj>I.^"H.gjhkMa30>'ٿH" ~V d`aWWO?\êwXQªxc+>O&-aU>zɥxA&IT4_g3>)uKKo$pPĈIvL>6cyT|L~ǐУBϟBHda O `D"| _ ;Q!ahDIHJܟj-OA8zS %\MD:Yg4֘$ƛOſoTsޙGo:A6b/ld-"iӛ?o66c[~i^}!?jo98xCvnnoܺʻDXU/d3֚u;a~X5[gmZV!\EtJ] J;fn#źEx=0uˇ獴nCh3W:`xl\I1w8:iC}Vu@ BɪԾR+X+;4^P5?Wdx "-EP~%JEMg 5kB rs萷J" !=06hwqz]b]9pIRrQ\DbQkxߨC0 Uǎ,U~_㸏 B +uQzU1s$8gܕ˞fT6qG;6AroSōNpT"*)uU$GI@Q,HDU(RW`,bѝ*vLe 4U3-]=LK tj~0~)U0Y n^f\ gbt]IiuD]7%c!/)(, p&-,)I(*v=$&RΟ"i?-?_f;Ir!'/{pgeſWWEw0\9vOe$r->?VtB,@JPNڹ3w8q~LY)[zo0L̾F9$@?=dX#ϗw>&vIKZ 6b4X4ZR2" %h萋q~9 EzdlpmP.S\"%ģt//AMIZ(tF42?'1zrsq2? E``SډDuLtRJ3&-.s؇8283X4sStqC0_x"tKlP^x|f1Ys~S}v._x0z^qȊ9t* )Z5ZoMj=QCX!^Dd|ZpW*Q{rDÙVDs)W 竞Km>Q8) |:i"?}ZIsy, iͰZȣfR|GC?jpjєJ{/ yo?,hR!_?\wzjχ! B~-*!+C,z_RSPQ[+b%:E߇|פ%ȟ>z x P]-vCPov^]Tk-4Q)y};I*(Keϴ{>$c+n xs[Ò0j5K$`jo~DKds,xr;}3X%HQ+5YhUDir ?WkY[o,j6YA N%\W'ZMJ|\ݙ}œǻ|z$SxכMWO$^$;&5"s`f|7y$X91@=urٻ6r&Wn"Yu^৭XIv߷(;(&LISd}LNP "̈ގ1^)i5+5z%kX(0CgrXQf2M~eesؐx;~^K"g'[И ~KVղ&+{~ X u)z/pV#P ߽3ʈe1Xo 3~F*]Ќ/?p {.{ :o`3R(b}bL%UL LGe'{%#~ qwW{[F|:-#>ǐzN#32#M|mX4ﵳlCfx, OlA9!{OpF{(cH\:~!q5ޱ` *f=T=.G, /'v18'/ט*'>|EFa*#sGGk/G6 /-GEF{X#Z+8OCD]W4[  o0k٥ȁwB1?ǂq t{KZU_3[o>Wbn̼]?g6nWvw!?w仫۫ve۴'vqKŧ^C@J".}Ps;x_/W쵥%B9y(" e =}elnۇLyeKZ)P]޽>vUAV)5&Z#阈(OΕŘ>SJok2*o,VN({[|+Mjag,wJQ̙I֋u1*$$DǻQDI@҉eQ%C?E'/K8Y*…d9,YC+FT^/ia^ϛ-O1mGX+"+#? #+~aQ'A[,B{y±A۫k~MgG0ٞT51)\Yh)ln@"R)Vq{V4`f_+J^ݰ/?bTOfp#}%em6(0J]ZjL<2Vj36He6_ް7/d4/I\],u{#]a,:ʚϢYZy4okr-YenmDkftI7.e,Mm"=e (d*L^+MT$ -Sal2x^E}n ShkMۼM[eQ1^OEv,e H`ٹۛ0H4O2dZ2f`5ܖ@ b Q -t)t>F` yX/^|+=ۉXMY55[j)^C pFikO py$lF2@@/|U (B.a54ɩ NVJaW [ŖL]T-lԺ,$.F)3tsD^O~ڟHef7>|O%`DDoZ< W uGO>B@Қy 0FHUgSGƑw e B97w5ZrN5K|JHƒ`󬞆8GpDҚJʮޢ*4 IJN'YQ73_8Zڕ"Z]҆[7 CYahd 6Z}X|K8Żz;H.e{Thn JPiR{hg?v۶1SJ:q+w5 ^!tf( #Cy^1$A{/Lr؆A[C/YfAXSXUB2X:FثƝN˻h]KS$`b WP*MdšJ/x,JO*lB6m-x+}M( &1Hk/ Uv0%0ځe]HbNV}6Tb)fKђ P\@;8v:<ӡ友sdE:8e^\[p +c+OjPE3Hڞ ~c]X')+ T}yf;Z;x * d cՙd:((#vPC Yp25oҘc6^)gk-X3Ym8"Ϳܩ+gls 5dZ|Q **RznX\{ISݢ ]q;}XF(2V|Qw s=ܯKzn8^Ma5nݷmqYY[3mEAjQlǾ\0"/(jφXӹBn4:q4ۦbu"]HDIdSo$I"TIMiW(U:ݷ#bSOio8vudJKkזR]C꜋RZN9F:'k?罫a` 2IWKh]r>OZO4b, HrrrQ/"cҦuXc+$W-pYa59#2Jat$%j(l!euBͨO9~{yHe'رlu%RXGTXxv {8|@L̀l{1HO5iGY:=cY7'U(5P)ەx>?WuxH. {5z 9$I )U>D'Tq:&vq%cVM]q&HU @HB:pZՇJ-nGCsv YVRwc) ԔWaX) LEZiR᳎vr?d]VK*s%fFd6썶UD R`S`t<6FL}w}FG#3z`کUPB'\dscciB&lbikwW[H6NX8y70Ҳ: h4?M.jQe!zGdz`)fTLI$4QAK=I= Fy%\ -<iY:0bR+;& B(E]-tƏʼn`{nu,1xW,D-D1(4FՅZQnpq;k]j;#XpkvfGjޚ߈Wl:7A{`q"sKg t֌%Fռ%꬀ݮrؾLD/dK3@#lv,#YGX~5[OZ,֏DpdZJ w>^G2Z3Z9"F +]kG=oe^Z!AarpSV}."a<a\ !i̶@wģ))s*ؖy[/%s)RHU帲 IX Ѻ v>m}XX>nٹMo#XMk> -?J )S7 YGu؏7L-gs. o@KA2A:& L)7A&(M " 2>![#BqR$$ Ӏh@4?{լ{O[s Z]x>{7TLѸU$/}!k\V) @050cdCc@qW;*-գ*!ar@)@V*Kau;Oz/Z w(e -LQ됶c))NJ>]t-{sR&l\!P$׺~1>Rb,{/K8z,~Oßfvh^̫>LF]]OQj>A{K)Ǭ\S@Kjә_v }#DBf|2C n3^0)k̔{Gdn H ܄!,cV([$gRV%d[+%,LGʃ]$/&(h@7kә3pƏ *l}q2yLiXՔ K(>'m˭Iv/މ`G?Wf(k R/wܢEw*.`H`a?!dx'Cx -B$K =5 0 l x&3:̽.M9B(by&RL2":ϭҢX!;tgKًؐ;$l un4a;ӰpFwm[bT>_;) Ή,BB 3LӍo76 84ɐ3>ٹaEB{ !S-JH׏6F!BXubIሥ h?|"gW' 7o}t.?0&LxE[VDahpYa9 ڝ#Ɓ G,ERqUB焳 /gR6{> }wپ-[x_8 qhsq:UThk0 9[^: 'C_bcx[`{JOjvNΰ_V4FtJwz`d!%z8R'I[,vp6 ..0e 9x8 o*&jH4\П߱ {??AL8B2^6 v-ܵm}om~!3X^?NSӎ B 59*<8$Z/H[k'S^!%88[+/qLAXDe`VEMG%T^#̛~{ie?^XVxqZ,`7&U+$_WXIb|S+ jv'iR ۏ|Y%{ӌSoJBPQ1R&5D@!"y'ɂfJ9 b8A{?[+pW߮{7}=ت A14C*5U]\34pŨ~~,Nzw<]ڃ׌{z .%AcX""޼\8FϹ^} C(vW!]/1hBUuXՊBkZqE[cQjD\TI;jQKsFoGnZQ@M(+oVr뒔;kx4UTef;(8%]eTԐy2iPgج"*'K",KXf΄#z 6JѪz<, qXVPU5;J?T !hX\"io&O/.}^p ?|&`bLR@,0I{I9 `0 碕(_^ -gcrIag&gk@]}|\|X6u𗥳; =.{ lX*A~?xN0ZoyuxE03`WXXw|iGɫΑB?J~XWOSo2(aoK(ЅB&}v2).s9D|Z!FB:`6Ĩ*;q9FcKj5Wzcs']ٴ̪/z[T` b/oaۢ|goq6SAk 4aR8XSg :YæT5DtdMe #QsY oW$`00<1id*MMȉ+ʰgC8&YMGxR^sJrz'Mg *' Uȩ?i=Z'Ԅ#\S!u@n!ߕ=T@Nk"R20ڱlJ:2yϑṒLXꔦ1r٧,VRBLƔ^ M$+#ۯ58vԨ:5jL?PT27jֲ"RCҨ̸@gTǜJ@gޠbJJj! ²S|<c`̝҇F8FWwAsŹ3})s 0I=c,`ci&謷;UqZ}ҁ[ Z9;"AɢM]Q@-FxǗ.ͥsuCk}^ٗ7V8WS[Du~g6 %6L#Cc8^@Rgfh1M1 /cgyS54~tca۝ދݮn}ABIi[j!E,D:iS3ʚ9cs0VhlGt#E7u1EĨ1547(";I FXq:Y+o-γN&HNAƚW)0heFr&Rd}~>!1*pBho *Bz[Gd@Ɏwew"dkr/G99Dr,Yd?۱XcԕzwO2:K5xr6F+,3:']. <'\F=۶TS&|!+b*B 6Ȇ$*haɄ$E4YzU|zwԣIc3 ޺C+Ѻߺ)e@xG`SFKvVFQpm݁2ae=kU^;VQwۢV00Xl(5gU˰4>Ohxp ű$OfO !7pT"detɁ=ƈ"A[pX"Q_n̲UFO\AAX-4i)e#1&dXt(z aTn`&vTY/ 5I&C=tM"V[P6j\ 4s FIN2k L쀀[C~&UɻmQAA5NkR5Ž s(v}`_mzDAR"wԷ("`5CݒM>@<@`t@e<4OMC9xPV3~{3ekq;f5N4\0&kM45c=1LA}4~}Dκ>UzdֈpM%.Hr%^ I &~%'V}99 ՗szq^a|w/}%T:zmRsq;-uW݇Vccm㝼|\;R %rjt= گ D1avJMZ|P_3XňUfu_wr r!PglNu8cy3ki[,Em ,`j`qHoۋ5Q;=//+P51F'ant5sЂw[ռ/Փ_+tHSP[~GP Mg/wuPub7X>\_}YNez 0kr~0ґp-)H!F7ިJec:hX̀N[%օp-C8FT=FRN;D}"jll;̌n]X 7ʼYCyBZog%{A+ze o]xU6mxŷl˗Xk1ZCk9pZTxgٝ|hC$Շ&L>GV=v²c(fCz 0ENMc} 8BB~\^^4d=6R\5mn pwdtQ#4nwZV\ i4ץĽOR euLUV']K+ zZl=IB^ؔa3覤03ua!/DKlJS~覛CJec:hЬkVpݺn+69ʾy?Ӛr&ɼsvk@HvVt4 h\! .r?]L/X0q!KD:VbxwEhT6@n{0? ݝ/ރs#B)*:^R.%Dryf37E.͑~qTXc+B60-ƴ{4CR~r|I)ZP%RFzbUjLp^ja vڛ$<@[Ra5╀_(BFPFrqUG$0P+߲a9 S!KZ KX8؊sHBȤ!zL&-բwPE ULEbWVT+Pƍ35BoY"!UU99:|"0`i=!&s1(8P(9BȴGI}; -x :SuE TCt,r'%>vc~*bZcLϐ/0IE]4VZ W0Y_\LCHfώ-a٠0  +g%DfJ+Nk)Ì<:+ᘫg04M0| =k"j mAW̞6HH?`Tgh|Xnɮo旬+F_>} ;%c 564xJR9UmJz[>M OB\dF 2QH䔔SNj.\H¶zW+Q~uQ3Ө"\!+4El<&{qZk '=?~?׳ApA4LmU|ZKN7v[pmK$LQi~[DQW xa x H`"zp&Z\-gpr9<*΍jd>(n%X˗cb/E9ۢ67Z*c0]~|8ΪH௕p42"H )e8^Mɤآ!O@IiZmb ;䚶d@EYI8[`av}Џ[0+"0+1Ÿ-΂îz!7Flx9\UhZ4((RV39k*+2v~7RlGF,ܼ#+E^^3X VJ10촤O9QR&9Qgbtׯz"LimS2KO^K?geN{5=Dir$FWoa{q6''Z}1c7sf78*'!^MӠzZA"oB椁Ys+^Ӭw5{ýMNf0Z=A@+ d (A򸲒qkyAj8ϔB5M].\23[.V͛+]Ɖ=Wr NF!)9>/OV~ڷy7,=3 /pBO DsX\[wI+N>\^԰.{cZnv<{8F)P ;)e,U m,)]7OPIs//Y6-Qt^9 wUb9E>Ѭ':{S4㍆YML ~4 |VnR3}0KCqfߺf^7k޽q<F_5fs̬nв2؜ J`Hqtc&A?'#XWC{ fϤshZ}VI?:CiMF#(>lMJ'ÛcA']ڜMM >BYL-#w&.~@_gN)Tp$*fLG4N31cCmtbxFqRXkD?K\AD+quk~%JM.f3+ Fqʱ'*|A0#\qGOE]U`uBxV}Kdn Wa·* [Oط 97I1ČTycq"DDD1L '+vN=ZK뜗, G &\jc|T;r4NW,R#bOyt>0A~3U'Kr?Ųss ,BȤZij~pGmb d^rT]ƽsP8 P7'^@iwog.'\+eޝoA$v\%r;!V6ZLt]\qA9)IG5 %DsrX$i{m.iǠ|']$9ژ}:20 7&9D^p݋^̼Wz1[ 0av:bї%{q"6_O/_ݢ59ԨƄ? Rc Q^lBJ6݋/fj#a b JmwF4ƳL\ Laqvo\S7-)0"Ū&.' ,&Bo1!˚]CjtYN{Znٱ.@SQĩ4y I^O4`u۴c+el;!^/3}Kdfj*4'[>kݗ-uEWrv_Lwsx}<Ӵ;XPϚw2: 0!{eߒh%k8凖;AAB0h?h8lх":pC.DFD 2JG'/lx&΋Cwum0_ ZGR% S}5,mR _?ٱ}>*2EDQ'HVZz08itʞSdlC5Lg*kx L)/OTX=R\B"R@SJ;O9x12&7H3% =Z$6-@Am/mqUUUiRcSh&u>~QCRuqxz/-_hWʯK~fß-ǯsYP%Ĕ[>lN_îew_| @eM/¿??o<= qFWo0AdL"Okpb`c>:C _\qa 0ŕ|< AW ']iwל :gݶWTAde );rͷ6Ck;,й/D%3C'i%N!tZ\>>^ZG>JL.GIqMH'\ E 5filjUax=9JP9iSGڽ;N[,jl}Mi^|&h!ӺvDc; 5u[):((c]C9JBr87eaQ6q6aH:,|[jf*/:4/Jci}J`(=$|[갅'*1hWSPv(UҥVQ ,i(]HM]&:j-](ZkDw6lk>bsrT5^JRD-I0hmDH?NC)i(=$|[j;F)IJiӃ@)IJR FiY/ɼ!ԀA4]M(#J(q\12z-5ߞP:bJbF,!J.6$ҸV5vAilWR1TBfJ",( y7a3䦂<kLحB#UL?YN'*Q+25hsb/%mEAaZ>@<^T㕘L[^\ 48 X+2J:'YZ)DzcC8׭@ vNYV4T"3f|%=3ABqBblX byNeD8ϙM˪?LKa'ALq1YemgB( . fC. 55ݥ \` j?JݧZf{~-O*OU pr$ _ĶF[g{~cb*R~dKYQeÝg"Oқ%c7F':c4AUt31d8PLVVKKX<%&t]t{ Gd\p5>st'ZC@5hw͙var/home/core/zuul-output/logs/kubelet.log0000644000000000000000004640770715156226167017727 0ustar rootrootMar 17 09:09:36 crc systemd[1]: Starting Kubernetes Kubelet... Mar 17 09:09:36 crc restorecon[4681]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 09:09:36 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 09:09:37 crc restorecon[4681]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 17 09:09:37 crc restorecon[4681]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 17 09:09:38 crc kubenswrapper[4813]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 09:09:38 crc kubenswrapper[4813]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 17 09:09:38 crc kubenswrapper[4813]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 09:09:38 crc kubenswrapper[4813]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 09:09:38 crc kubenswrapper[4813]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 17 09:09:38 crc kubenswrapper[4813]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.398251 4813 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412511 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412560 4813 feature_gate.go:330] unrecognized feature gate: Example Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412569 4813 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412577 4813 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412586 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412622 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412631 4813 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412640 4813 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412648 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412656 4813 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412664 4813 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412672 4813 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412679 4813 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412687 4813 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412695 4813 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412702 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412710 4813 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412717 4813 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412725 4813 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412735 4813 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412745 4813 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412759 4813 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412767 4813 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412775 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412783 4813 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412791 4813 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412799 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412806 4813 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412814 4813 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412821 4813 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412829 4813 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412837 4813 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412845 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412855 4813 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412866 4813 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412875 4813 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412885 4813 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412893 4813 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412901 4813 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412909 4813 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412916 4813 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412924 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412931 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412939 4813 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412947 4813 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412955 4813 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412963 4813 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412971 4813 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412979 4813 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412987 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.412998 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.413006 4813 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.413018 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.413026 4813 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.413034 4813 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.413042 4813 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.413050 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.413060 4813 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.413073 4813 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.413081 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.413089 4813 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.413097 4813 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.413105 4813 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.413113 4813 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.413121 4813 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.413128 4813 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.413135 4813 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.413150 4813 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.413161 4813 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.413171 4813 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.413180 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413326 4813 flags.go:64] FLAG: --address="0.0.0.0" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413344 4813 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413363 4813 flags.go:64] FLAG: --anonymous-auth="true" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413378 4813 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413393 4813 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413404 4813 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413418 4813 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413433 4813 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413445 4813 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413457 4813 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413469 4813 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413484 4813 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413496 4813 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413507 4813 flags.go:64] FLAG: --cgroup-root="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413517 4813 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413528 4813 flags.go:64] FLAG: --client-ca-file="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413537 4813 flags.go:64] FLAG: --cloud-config="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413546 4813 flags.go:64] FLAG: --cloud-provider="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413555 4813 flags.go:64] FLAG: --cluster-dns="[]" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413565 4813 flags.go:64] FLAG: --cluster-domain="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413574 4813 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413583 4813 flags.go:64] FLAG: --config-dir="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413592 4813 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413630 4813 flags.go:64] FLAG: --container-log-max-files="5" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413642 4813 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413651 4813 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413661 4813 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413670 4813 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413679 4813 flags.go:64] FLAG: --contention-profiling="false" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413689 4813 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413698 4813 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413709 4813 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413718 4813 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413735 4813 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413744 4813 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413753 4813 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413762 4813 flags.go:64] FLAG: --enable-load-reader="false" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413770 4813 flags.go:64] FLAG: --enable-server="true" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413779 4813 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413794 4813 flags.go:64] FLAG: --event-burst="100" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413804 4813 flags.go:64] FLAG: --event-qps="50" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413813 4813 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413822 4813 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413831 4813 flags.go:64] FLAG: --eviction-hard="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413841 4813 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413850 4813 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413859 4813 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413869 4813 flags.go:64] FLAG: --eviction-soft="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413878 4813 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413887 4813 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413896 4813 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413906 4813 flags.go:64] FLAG: --experimental-mounter-path="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413915 4813 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413924 4813 flags.go:64] FLAG: --fail-swap-on="true" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413933 4813 flags.go:64] FLAG: --feature-gates="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413952 4813 flags.go:64] FLAG: --file-check-frequency="20s" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413962 4813 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413971 4813 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413980 4813 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413990 4813 flags.go:64] FLAG: --healthz-port="10248" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.413999 4813 flags.go:64] FLAG: --help="false" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414008 4813 flags.go:64] FLAG: --hostname-override="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414017 4813 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414027 4813 flags.go:64] FLAG: --http-check-frequency="20s" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414036 4813 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414045 4813 flags.go:64] FLAG: --image-credential-provider-config="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414054 4813 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414064 4813 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414073 4813 flags.go:64] FLAG: --image-service-endpoint="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414082 4813 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414092 4813 flags.go:64] FLAG: --kube-api-burst="100" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414101 4813 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414110 4813 flags.go:64] FLAG: --kube-api-qps="50" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414119 4813 flags.go:64] FLAG: --kube-reserved="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414128 4813 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414137 4813 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414146 4813 flags.go:64] FLAG: --kubelet-cgroups="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414155 4813 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414164 4813 flags.go:64] FLAG: --lock-file="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414172 4813 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414182 4813 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414190 4813 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414203 4813 flags.go:64] FLAG: --log-json-split-stream="false" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414214 4813 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414223 4813 flags.go:64] FLAG: --log-text-split-stream="false" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414232 4813 flags.go:64] FLAG: --logging-format="text" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414241 4813 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414250 4813 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414259 4813 flags.go:64] FLAG: --manifest-url="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414268 4813 flags.go:64] FLAG: --manifest-url-header="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414280 4813 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414289 4813 flags.go:64] FLAG: --max-open-files="1000000" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414299 4813 flags.go:64] FLAG: --max-pods="110" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414308 4813 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414318 4813 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414327 4813 flags.go:64] FLAG: --memory-manager-policy="None" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414336 4813 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414345 4813 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414354 4813 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414364 4813 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414383 4813 flags.go:64] FLAG: --node-status-max-images="50" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414393 4813 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414402 4813 flags.go:64] FLAG: --oom-score-adj="-999" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414411 4813 flags.go:64] FLAG: --pod-cidr="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414420 4813 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414433 4813 flags.go:64] FLAG: --pod-manifest-path="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414442 4813 flags.go:64] FLAG: --pod-max-pids="-1" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414452 4813 flags.go:64] FLAG: --pods-per-core="0" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414461 4813 flags.go:64] FLAG: --port="10250" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414470 4813 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414479 4813 flags.go:64] FLAG: --provider-id="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414488 4813 flags.go:64] FLAG: --qos-reserved="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414497 4813 flags.go:64] FLAG: --read-only-port="10255" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414506 4813 flags.go:64] FLAG: --register-node="true" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414515 4813 flags.go:64] FLAG: --register-schedulable="true" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414524 4813 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414539 4813 flags.go:64] FLAG: --registry-burst="10" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414548 4813 flags.go:64] FLAG: --registry-qps="5" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414556 4813 flags.go:64] FLAG: --reserved-cpus="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414566 4813 flags.go:64] FLAG: --reserved-memory="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414577 4813 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414587 4813 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414617 4813 flags.go:64] FLAG: --rotate-certificates="false" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414627 4813 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414636 4813 flags.go:64] FLAG: --runonce="false" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414644 4813 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414654 4813 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414664 4813 flags.go:64] FLAG: --seccomp-default="false" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414672 4813 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414681 4813 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414690 4813 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414700 4813 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414709 4813 flags.go:64] FLAG: --storage-driver-password="root" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414718 4813 flags.go:64] FLAG: --storage-driver-secure="false" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414727 4813 flags.go:64] FLAG: --storage-driver-table="stats" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414736 4813 flags.go:64] FLAG: --storage-driver-user="root" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414744 4813 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414754 4813 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414764 4813 flags.go:64] FLAG: --system-cgroups="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414773 4813 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414787 4813 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414796 4813 flags.go:64] FLAG: --tls-cert-file="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414805 4813 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414815 4813 flags.go:64] FLAG: --tls-min-version="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414824 4813 flags.go:64] FLAG: --tls-private-key-file="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414832 4813 flags.go:64] FLAG: --topology-manager-policy="none" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414842 4813 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414851 4813 flags.go:64] FLAG: --topology-manager-scope="container" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414860 4813 flags.go:64] FLAG: --v="2" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414871 4813 flags.go:64] FLAG: --version="false" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414883 4813 flags.go:64] FLAG: --vmodule="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414893 4813 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.414903 4813 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415106 4813 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415117 4813 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415127 4813 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415136 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415144 4813 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415153 4813 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415162 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415171 4813 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415179 4813 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415187 4813 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415195 4813 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415203 4813 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415210 4813 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415218 4813 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415227 4813 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415237 4813 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415246 4813 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415256 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415266 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415274 4813 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415283 4813 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415290 4813 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415298 4813 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415306 4813 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415314 4813 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415322 4813 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415330 4813 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415338 4813 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415346 4813 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415353 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415361 4813 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415370 4813 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415377 4813 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415385 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415392 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415400 4813 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415408 4813 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415422 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415433 4813 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415442 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415453 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415462 4813 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415473 4813 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415483 4813 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415493 4813 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415502 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415511 4813 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415546 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415556 4813 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415565 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415573 4813 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415581 4813 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415590 4813 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415620 4813 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415628 4813 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415636 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415644 4813 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415655 4813 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415665 4813 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415674 4813 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415683 4813 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415691 4813 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415700 4813 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415709 4813 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415716 4813 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415724 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415732 4813 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415740 4813 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415748 4813 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415758 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.415767 4813 feature_gate.go:330] unrecognized feature gate: Example Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.415781 4813 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.429952 4813 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.429992 4813 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430116 4813 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430130 4813 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430141 4813 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430150 4813 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430159 4813 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430167 4813 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430175 4813 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430184 4813 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430193 4813 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430202 4813 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430209 4813 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430217 4813 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430225 4813 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430233 4813 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430241 4813 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430249 4813 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430257 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430265 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430274 4813 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430282 4813 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430290 4813 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430298 4813 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430306 4813 feature_gate.go:330] unrecognized feature gate: Example Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430315 4813 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430323 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430331 4813 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430339 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430347 4813 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430354 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430363 4813 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430370 4813 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430378 4813 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430386 4813 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430393 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430414 4813 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430423 4813 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430433 4813 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430443 4813 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430452 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430460 4813 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430471 4813 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430481 4813 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430491 4813 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430498 4813 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430506 4813 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430514 4813 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430522 4813 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430530 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430538 4813 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430546 4813 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430553 4813 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430564 4813 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430573 4813 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430581 4813 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430588 4813 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430620 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430628 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430635 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430643 4813 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430651 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430659 4813 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430668 4813 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430678 4813 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430688 4813 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430696 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430707 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430715 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430723 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430731 4813 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430739 4813 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.430756 4813 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.430769 4813 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431018 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431034 4813 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431043 4813 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431052 4813 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431060 4813 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431068 4813 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431075 4813 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431083 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431091 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431098 4813 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431106 4813 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431114 4813 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431122 4813 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431133 4813 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431143 4813 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431151 4813 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431160 4813 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431168 4813 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431176 4813 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431184 4813 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431191 4813 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431199 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431207 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431216 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431225 4813 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431234 4813 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431243 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431250 4813 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431258 4813 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431265 4813 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431273 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431280 4813 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431288 4813 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431295 4813 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431304 4813 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431312 4813 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431320 4813 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431328 4813 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431336 4813 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431344 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431352 4813 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431360 4813 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431368 4813 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431378 4813 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431385 4813 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431393 4813 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431401 4813 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431409 4813 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431416 4813 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431424 4813 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431434 4813 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431444 4813 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431453 4813 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431464 4813 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431473 4813 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431484 4813 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431494 4813 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431504 4813 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431513 4813 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431521 4813 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431550 4813 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431558 4813 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431566 4813 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431574 4813 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431581 4813 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431589 4813 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431622 4813 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431636 4813 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431658 4813 feature_gate.go:330] unrecognized feature gate: Example Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431669 4813 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.431683 4813 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.431697 4813 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.431949 4813 server.go:940] "Client rotation is on, will bootstrap in background" Mar 17 09:09:38 crc kubenswrapper[4813]: E0317 09:09:38.436626 4813 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.443114 4813 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.443318 4813 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.446219 4813 server.go:997] "Starting client certificate rotation" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.446282 4813 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.446583 4813 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.491264 4813 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.494091 4813 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 17 09:09:38 crc kubenswrapper[4813]: E0317 09:09:38.494285 4813 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.143:6443: connect: connection refused" logger="UnhandledError" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.510694 4813 log.go:25] "Validated CRI v1 runtime API" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.564849 4813 log.go:25] "Validated CRI v1 image API" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.567292 4813 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.573940 4813 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-17-09-05-12-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.573989 4813 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.604536 4813 manager.go:217] Machine: {Timestamp:2026-03-17 09:09:38.598892166 +0000 UTC m=+0.699695705 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:63a6017f-37e1-4eb4-b66c-a7f05e6608e4 BootID:ea32862a-5b23-419b-aecf-996f0626022b Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:7a:eb:29 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:7a:eb:29 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:f9:36:08 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:2f:47:68 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:dc:5a:06 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:4e:81:45 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:3a:86:b5:11:a4:23 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:d6:ec:8b:b8:a5:78 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.604998 4813 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.605196 4813 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.606724 4813 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.607083 4813 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.607145 4813 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.607502 4813 topology_manager.go:138] "Creating topology manager with none policy" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.607529 4813 container_manager_linux.go:303] "Creating device plugin manager" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.608236 4813 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.608300 4813 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.609315 4813 state_mem.go:36] "Initialized new in-memory state store" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.609479 4813 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.614747 4813 kubelet.go:418] "Attempting to sync node with API server" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.614788 4813 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.614895 4813 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.614927 4813 kubelet.go:324] "Adding apiserver pod source" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.614950 4813 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.627542 4813 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.630070 4813 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.631568 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.143:6443: connect: connection refused Mar 17 09:09:38 crc kubenswrapper[4813]: E0317 09:09:38.631744 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.143:6443: connect: connection refused" logger="UnhandledError" Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.632396 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.143:6443: connect: connection refused Mar 17 09:09:38 crc kubenswrapper[4813]: E0317 09:09:38.632475 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.143:6443: connect: connection refused" logger="UnhandledError" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.633241 4813 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.635567 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.635677 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.635692 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.635706 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.635733 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.635751 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.635767 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.635795 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.635817 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.635834 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.635858 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.635875 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.647578 4813 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.648304 4813 server.go:1280] "Started kubelet" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.653999 4813 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.653997 4813 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 17 09:09:38 crc systemd[1]: Started Kubernetes Kubelet. Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.656859 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.143:6443: connect: connection refused Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.657468 4813 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.659560 4813 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.659693 4813 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.660141 4813 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 17 09:09:38 crc kubenswrapper[4813]: E0317 09:09:38.660134 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.660817 4813 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.662136 4813 factory.go:55] Registering systemd factory Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.662180 4813 factory.go:221] Registration of the systemd container factory successfully Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.664478 4813 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.664981 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.143:6443: connect: connection refused Mar 17 09:09:38 crc kubenswrapper[4813]: E0317 09:09:38.665110 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.143:6443: connect: connection refused" logger="UnhandledError" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.665050 4813 server.go:460] "Adding debug handlers to kubelet server" Mar 17 09:09:38 crc kubenswrapper[4813]: E0317 09:09:38.665319 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.143:6443: connect: connection refused" interval="200ms" Mar 17 09:09:38 crc kubenswrapper[4813]: E0317 09:09:38.664724 4813 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.143:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189d95d6f0bc735a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.648265562 +0000 UTC m=+0.749069101,LastTimestamp:2026-03-17 09:09:38.648265562 +0000 UTC m=+0.749069101,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.667326 4813 factory.go:153] Registering CRI-O factory Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.667465 4813 factory.go:221] Registration of the crio container factory successfully Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.667798 4813 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.667953 4813 factory.go:103] Registering Raw factory Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.668075 4813 manager.go:1196] Started watching for new ooms in manager Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.669791 4813 manager.go:319] Starting recovery of all containers Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678287 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678370 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678401 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678427 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678451 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678479 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678505 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678531 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678560 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678584 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678642 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678670 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678694 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678725 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678749 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678778 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678802 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678827 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678851 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678880 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678905 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678932 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678958 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.678983 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679008 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679031 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679064 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679093 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679119 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679144 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679169 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679198 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679223 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679248 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679272 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679295 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679319 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679342 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679367 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679392 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679415 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679439 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679463 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679489 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679515 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679543 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679567 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679593 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679695 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679728 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679753 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679777 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679812 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679839 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679866 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679891 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679916 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679943 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679966 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.679991 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680013 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680038 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680062 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680088 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680112 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680139 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680166 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680189 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680217 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680244 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680267 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680293 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680316 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680338 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680361 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680385 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680409 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680433 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680458 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680486 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680511 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680538 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680561 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680589 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680818 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.680846 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684357 4813 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684406 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684426 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684440 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684455 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684469 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684482 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684496 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684510 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684525 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684540 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684556 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684569 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684582 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684614 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684628 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684641 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684656 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684669 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684691 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684712 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684728 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684744 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684759 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684775 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684795 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684810 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684824 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684839 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684852 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684865 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684879 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684892 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684905 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684917 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684932 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684968 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684982 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.684998 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685012 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685026 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685048 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685061 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685075 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685089 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685102 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685116 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685128 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685142 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685155 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685168 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685181 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685194 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685207 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685223 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685239 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685253 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685267 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685281 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685295 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685312 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685562 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685583 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685620 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685635 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685648 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685663 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685676 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685787 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685802 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685815 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685828 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685850 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685864 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685877 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685888 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685903 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685917 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685930 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685941 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685954 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685966 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685979 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.685993 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686005 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686019 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686033 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686045 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686059 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686072 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686085 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686098 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686110 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686122 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686134 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686148 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686160 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686173 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686185 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686199 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686211 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686224 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686238 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686251 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686265 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686280 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686292 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686304 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686319 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686333 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686344 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686357 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686370 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686382 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686394 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686407 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686420 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686434 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686447 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686459 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686472 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686485 4813 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686498 4813 reconstruct.go:97] "Volume reconstruction finished" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.686506 4813 reconciler.go:26] "Reconciler: start to sync state" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.702774 4813 manager.go:324] Recovery completed Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.719784 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.722065 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.722105 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.722118 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.723064 4813 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.723082 4813 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.723102 4813 state_mem.go:36] "Initialized new in-memory state store" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.724135 4813 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.729287 4813 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.729346 4813 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.729382 4813 kubelet.go:2335] "Starting kubelet main sync loop" Mar 17 09:09:38 crc kubenswrapper[4813]: E0317 09:09:38.729464 4813 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 17 09:09:38 crc kubenswrapper[4813]: W0317 09:09:38.730506 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.143:6443: connect: connection refused Mar 17 09:09:38 crc kubenswrapper[4813]: E0317 09:09:38.730622 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.143:6443: connect: connection refused" logger="UnhandledError" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.744749 4813 policy_none.go:49] "None policy: Start" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.745425 4813 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.745472 4813 state_mem.go:35] "Initializing new in-memory state store" Mar 17 09:09:38 crc kubenswrapper[4813]: E0317 09:09:38.761733 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.811358 4813 manager.go:334] "Starting Device Plugin manager" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.811427 4813 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.811445 4813 server.go:79] "Starting device plugin registration server" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.812179 4813 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.812202 4813 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.812463 4813 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.812573 4813 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.812583 4813 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 17 09:09:38 crc kubenswrapper[4813]: E0317 09:09:38.819383 4813 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.829807 4813 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.829933 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.831140 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.831218 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.831233 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.831348 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.831661 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.831716 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.832068 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.832101 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.832111 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.832218 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.832443 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.832507 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.832529 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.832509 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.832541 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.832873 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.832891 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.832900 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.832993 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.833075 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.833124 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.833501 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.833540 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.833552 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.834829 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.834870 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.834886 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.835023 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.835048 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.835060 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.835077 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.835256 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.835288 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.836266 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.836289 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.836301 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.836506 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.836540 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.837033 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.837086 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.837106 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.838528 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.838556 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.838577 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:38 crc kubenswrapper[4813]: E0317 09:09:38.866751 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.143:6443: connect: connection refused" interval="400ms" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.890375 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.890440 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.890506 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.890581 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.890681 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.891311 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.891391 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.891432 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.891502 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.891533 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.891563 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.891594 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.891717 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.891748 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.891824 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.912689 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.914571 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.914666 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.914681 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.914717 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 09:09:38 crc kubenswrapper[4813]: E0317 09:09:38.915303 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.143:6443: connect: connection refused" node="crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.992732 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.992795 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.992818 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.992836 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.992853 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.992868 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.992885 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.992901 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.992921 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.992937 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.992953 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.992967 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.992968 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.992988 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.993022 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.992982 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.993042 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.993108 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.993124 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.993143 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.993159 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.993197 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.993198 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.993162 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.993225 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.993219 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.993282 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.993325 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.993165 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 17 09:09:38 crc kubenswrapper[4813]: I0317 09:09:38.993640 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 09:09:39 crc kubenswrapper[4813]: I0317 09:09:39.116437 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:39 crc kubenswrapper[4813]: I0317 09:09:39.118355 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:39 crc kubenswrapper[4813]: I0317 09:09:39.118422 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:39 crc kubenswrapper[4813]: I0317 09:09:39.118439 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:39 crc kubenswrapper[4813]: I0317 09:09:39.118480 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 09:09:39 crc kubenswrapper[4813]: E0317 09:09:39.119373 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.143:6443: connect: connection refused" node="crc" Mar 17 09:09:39 crc kubenswrapper[4813]: I0317 09:09:39.192722 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:09:39 crc kubenswrapper[4813]: I0317 09:09:39.220786 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 09:09:39 crc kubenswrapper[4813]: I0317 09:09:39.232268 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 17 09:09:39 crc kubenswrapper[4813]: W0317 09:09:39.244702 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-5f9be564cedd8ae9cf7fbe5bd09ba8beded1c53cebaa065c2fdf0817656d39cb WatchSource:0}: Error finding container 5f9be564cedd8ae9cf7fbe5bd09ba8beded1c53cebaa065c2fdf0817656d39cb: Status 404 returned error can't find the container with id 5f9be564cedd8ae9cf7fbe5bd09ba8beded1c53cebaa065c2fdf0817656d39cb Mar 17 09:09:39 crc kubenswrapper[4813]: I0317 09:09:39.256108 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 17 09:09:39 crc kubenswrapper[4813]: W0317 09:09:39.262517 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-8dcf1231670de91428d7eece3351f77a89d250a572257eef270dc96848240853 WatchSource:0}: Error finding container 8dcf1231670de91428d7eece3351f77a89d250a572257eef270dc96848240853: Status 404 returned error can't find the container with id 8dcf1231670de91428d7eece3351f77a89d250a572257eef270dc96848240853 Mar 17 09:09:39 crc kubenswrapper[4813]: W0317 09:09:39.265738 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-18b6ea730ac8f5117e09018d9a6b96cf6ad96db25bf6df0f124a1b8c88e46dd7 WatchSource:0}: Error finding container 18b6ea730ac8f5117e09018d9a6b96cf6ad96db25bf6df0f124a1b8c88e46dd7: Status 404 returned error can't find the container with id 18b6ea730ac8f5117e09018d9a6b96cf6ad96db25bf6df0f124a1b8c88e46dd7 Mar 17 09:09:39 crc kubenswrapper[4813]: I0317 09:09:39.266814 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:09:39 crc kubenswrapper[4813]: E0317 09:09:39.267533 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.143:6443: connect: connection refused" interval="800ms" Mar 17 09:09:39 crc kubenswrapper[4813]: W0317 09:09:39.283113 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-c39cde9fa3ed903e783124af74e085a3fa8c8445a233e4dbcd549719a7f55bb2 WatchSource:0}: Error finding container c39cde9fa3ed903e783124af74e085a3fa8c8445a233e4dbcd549719a7f55bb2: Status 404 returned error can't find the container with id c39cde9fa3ed903e783124af74e085a3fa8c8445a233e4dbcd549719a7f55bb2 Mar 17 09:09:39 crc kubenswrapper[4813]: W0317 09:09:39.293299 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-4cf35e9ea0c02bf9af73898d57d8db56d11086bfa7868a1da5d838395b49f12b WatchSource:0}: Error finding container 4cf35e9ea0c02bf9af73898d57d8db56d11086bfa7868a1da5d838395b49f12b: Status 404 returned error can't find the container with id 4cf35e9ea0c02bf9af73898d57d8db56d11086bfa7868a1da5d838395b49f12b Mar 17 09:09:39 crc kubenswrapper[4813]: I0317 09:09:39.519765 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:39 crc kubenswrapper[4813]: I0317 09:09:39.521355 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:39 crc kubenswrapper[4813]: I0317 09:09:39.521390 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:39 crc kubenswrapper[4813]: I0317 09:09:39.521399 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:39 crc kubenswrapper[4813]: I0317 09:09:39.521426 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 09:09:39 crc kubenswrapper[4813]: E0317 09:09:39.521747 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.143:6443: connect: connection refused" node="crc" Mar 17 09:09:39 crc kubenswrapper[4813]: W0317 09:09:39.553702 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.143:6443: connect: connection refused Mar 17 09:09:39 crc kubenswrapper[4813]: E0317 09:09:39.553832 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.143:6443: connect: connection refused" logger="UnhandledError" Mar 17 09:09:39 crc kubenswrapper[4813]: I0317 09:09:39.658388 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.143:6443: connect: connection refused Mar 17 09:09:39 crc kubenswrapper[4813]: I0317 09:09:39.743859 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"8dcf1231670de91428d7eece3351f77a89d250a572257eef270dc96848240853"} Mar 17 09:09:39 crc kubenswrapper[4813]: I0317 09:09:39.745257 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"5f9be564cedd8ae9cf7fbe5bd09ba8beded1c53cebaa065c2fdf0817656d39cb"} Mar 17 09:09:39 crc kubenswrapper[4813]: I0317 09:09:39.746499 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"4cf35e9ea0c02bf9af73898d57d8db56d11086bfa7868a1da5d838395b49f12b"} Mar 17 09:09:39 crc kubenswrapper[4813]: I0317 09:09:39.747388 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c39cde9fa3ed903e783124af74e085a3fa8c8445a233e4dbcd549719a7f55bb2"} Mar 17 09:09:39 crc kubenswrapper[4813]: I0317 09:09:39.748861 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"18b6ea730ac8f5117e09018d9a6b96cf6ad96db25bf6df0f124a1b8c88e46dd7"} Mar 17 09:09:39 crc kubenswrapper[4813]: W0317 09:09:39.902078 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.143:6443: connect: connection refused Mar 17 09:09:39 crc kubenswrapper[4813]: E0317 09:09:39.902185 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.143:6443: connect: connection refused" logger="UnhandledError" Mar 17 09:09:39 crc kubenswrapper[4813]: W0317 09:09:39.992255 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.143:6443: connect: connection refused Mar 17 09:09:39 crc kubenswrapper[4813]: E0317 09:09:39.992366 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.143:6443: connect: connection refused" logger="UnhandledError" Mar 17 09:09:40 crc kubenswrapper[4813]: E0317 09:09:40.068815 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.143:6443: connect: connection refused" interval="1.6s" Mar 17 09:09:40 crc kubenswrapper[4813]: W0317 09:09:40.218886 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.143:6443: connect: connection refused Mar 17 09:09:40 crc kubenswrapper[4813]: E0317 09:09:40.219623 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.143:6443: connect: connection refused" logger="UnhandledError" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.322644 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.325926 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.325960 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.325974 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.326001 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 09:09:40 crc kubenswrapper[4813]: E0317 09:09:40.326413 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.143:6443: connect: connection refused" node="crc" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.624530 4813 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 17 09:09:40 crc kubenswrapper[4813]: E0317 09:09:40.625521 4813 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.143:6443: connect: connection refused" logger="UnhandledError" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.658433 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.143:6443: connect: connection refused Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.754156 4813 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f" exitCode=0 Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.754254 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f"} Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.754800 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.755814 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.755855 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.755870 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.759014 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"3147869c6b9a5426228139c412d8b7b2cb224a771ca092fbe9fe3b9b2b215d61"} Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.759161 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.759164 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"faa97f1c57fb3a119266cee961a8872c9115b7541542c57598511fd47ea3c907"} Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.759403 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6a5b5d0590fefa483bcb5d5a78dfee7d61881fe9e19e5f229d3788d2cc09ab87"} Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.759424 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ee1d7eaebd9e22a3f68b00c57f0bc9971826fc05a41062947a1ce04f1e8d380c"} Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.760896 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.760953 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.760973 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.762879 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb" exitCode=0 Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.762987 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb"} Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.763014 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.764776 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.765281 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.765303 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.765459 4813 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="bcf270cfbb1e9f689f81743b5f68fd2ad3682e0891da1a27a0727f3c7840eb6d" exitCode=0 Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.765577 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"bcf270cfbb1e9f689f81743b5f68fd2ad3682e0891da1a27a0727f3c7840eb6d"} Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.765621 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.766878 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.766914 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.767038 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.767763 4813 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="03f72a274657c1a544b773168b70bde00e45c5eae7dde99271f43ce5ad87db72" exitCode=0 Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.767793 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.767781 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"03f72a274657c1a544b773168b70bde00e45c5eae7dde99271f43ce5ad87db72"} Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.769142 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.769206 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.769238 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.769248 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.770705 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.770749 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:40 crc kubenswrapper[4813]: I0317 09:09:40.770763 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:41 crc kubenswrapper[4813]: W0317 09:09:41.657257 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.143:6443: connect: connection refused Mar 17 09:09:41 crc kubenswrapper[4813]: E0317 09:09:41.657389 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.143:6443: connect: connection refused" logger="UnhandledError" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.657642 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.143:6443: connect: connection refused Mar 17 09:09:41 crc kubenswrapper[4813]: E0317 09:09:41.669849 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.143:6443: connect: connection refused" interval="3.2s" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.773634 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281"} Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.773709 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d"} Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.773732 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab"} Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.777053 4813 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="f49c003e0f99daac6c9359b311a0182cd6b638a911503d31bde4c04581f89c99" exitCode=0 Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.777144 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"f49c003e0f99daac6c9359b311a0182cd6b638a911503d31bde4c04581f89c99"} Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.777330 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.779077 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.779127 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.779140 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.781814 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"342e8bf02207ce28830cf7d394c4193c7c6722454114ba9febcf14cdd62d7ae5"} Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.781917 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.783322 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.783409 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.783437 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.787753 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.787806 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa"} Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.787861 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16"} Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.787882 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c"} Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.787891 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.789074 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.789117 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.789129 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.789503 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.789553 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.789570 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.873697 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.926879 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.941347 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.941391 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.941405 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:41 crc kubenswrapper[4813]: I0317 09:09:41.941438 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 09:09:41 crc kubenswrapper[4813]: E0317 09:09:41.941966 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.143:6443: connect: connection refused" node="crc" Mar 17 09:09:42 crc kubenswrapper[4813]: W0317 09:09:42.487174 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.143:6443: connect: connection refused Mar 17 09:09:42 crc kubenswrapper[4813]: E0317 09:09:42.487278 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.143:6443: connect: connection refused" logger="UnhandledError" Mar 17 09:09:42 crc kubenswrapper[4813]: W0317 09:09:42.533444 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.143:6443: connect: connection refused Mar 17 09:09:42 crc kubenswrapper[4813]: E0317 09:09:42.533570 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.143:6443: connect: connection refused" logger="UnhandledError" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.658231 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.143:6443: connect: connection refused Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.793734 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.796742 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="38fa2834a5940c4d76b73ca7a545601a5344a74c4bd8f5c77f02cf6044d84a0d" exitCode=255 Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.796878 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba"} Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.796961 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.796966 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"38fa2834a5940c4d76b73ca7a545601a5344a74c4bd8f5c77f02cf6044d84a0d"} Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.798481 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.798528 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.798538 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.799247 4813 scope.go:117] "RemoveContainer" containerID="38fa2834a5940c4d76b73ca7a545601a5344a74c4bd8f5c77f02cf6044d84a0d" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.802233 4813 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="fed1113da1f1a03a32776b473f5e8c699710f4d4b52e0ba795fc5eab447acb49" exitCode=0 Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.802295 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"fed1113da1f1a03a32776b473f5e8c699710f4d4b52e0ba795fc5eab447acb49"} Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.802437 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.802462 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.802514 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.802445 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.802449 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.804359 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.804392 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.804411 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.804999 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.805026 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.805036 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.805096 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.805137 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.805156 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.805151 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.805200 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:42 crc kubenswrapper[4813]: I0317 09:09:42.805227 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:43 crc kubenswrapper[4813]: I0317 09:09:43.725721 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 09:09:43 crc kubenswrapper[4813]: I0317 09:09:43.808457 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 17 09:09:43 crc kubenswrapper[4813]: I0317 09:09:43.821571 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5ceeff48801f75c62e01ddcf92c4a850d85cf514063f217b6cfd9c0294545b70"} Mar 17 09:09:43 crc kubenswrapper[4813]: I0317 09:09:43.821834 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:43 crc kubenswrapper[4813]: I0317 09:09:43.822521 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:09:43 crc kubenswrapper[4813]: I0317 09:09:43.823472 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:43 crc kubenswrapper[4813]: I0317 09:09:43.823517 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:43 crc kubenswrapper[4813]: I0317 09:09:43.823534 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:43 crc kubenswrapper[4813]: I0317 09:09:43.828411 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:43 crc kubenswrapper[4813]: I0317 09:09:43.828777 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4e6d4e31dbb0ec0ee02dd81bd2d654228c3229c7b07de90462c657ee97e45bb1"} Mar 17 09:09:43 crc kubenswrapper[4813]: I0317 09:09:43.828831 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"4b3e4ffb3c9572e4d93aaf170d629b12a806dd4be0eaeb5677ea33e1583b84d6"} Mar 17 09:09:43 crc kubenswrapper[4813]: I0317 09:09:43.828854 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"be5b6cc5bbd218b71befdd26e2f4f660a8e71a5e85caf5e6e9e36f0c3762ca4a"} Mar 17 09:09:43 crc kubenswrapper[4813]: I0317 09:09:43.828872 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bbce1d32f7663982ef5c5e5df60dc5a0045f508046991c42a44fad48c7794237"} Mar 17 09:09:43 crc kubenswrapper[4813]: I0317 09:09:43.830250 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:43 crc kubenswrapper[4813]: I0317 09:09:43.830322 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:43 crc kubenswrapper[4813]: I0317 09:09:43.830342 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:43 crc kubenswrapper[4813]: I0317 09:09:43.996547 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:09:43 crc kubenswrapper[4813]: I0317 09:09:43.996847 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:43 crc kubenswrapper[4813]: I0317 09:09:43.998343 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:43 crc kubenswrapper[4813]: I0317 09:09:43.998389 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:43 crc kubenswrapper[4813]: I0317 09:09:43.998405 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:44 crc kubenswrapper[4813]: I0317 09:09:44.342920 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:09:44 crc kubenswrapper[4813]: I0317 09:09:44.838664 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"e57df849184fd507ce39a1b31b6bef4da5c6c0f9aac69491cadfbc2926e6e0b8"} Mar 17 09:09:44 crc kubenswrapper[4813]: I0317 09:09:44.838773 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:44 crc kubenswrapper[4813]: I0317 09:09:44.838860 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:09:44 crc kubenswrapper[4813]: I0317 09:09:44.838777 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:44 crc kubenswrapper[4813]: I0317 09:09:44.840144 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:44 crc kubenswrapper[4813]: I0317 09:09:44.840193 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:44 crc kubenswrapper[4813]: I0317 09:09:44.840210 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:44 crc kubenswrapper[4813]: I0317 09:09:44.841433 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:44 crc kubenswrapper[4813]: I0317 09:09:44.841515 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:44 crc kubenswrapper[4813]: I0317 09:09:44.841542 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:44 crc kubenswrapper[4813]: I0317 09:09:44.865710 4813 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 17 09:09:45 crc kubenswrapper[4813]: I0317 09:09:45.142666 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:45 crc kubenswrapper[4813]: I0317 09:09:45.144167 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:45 crc kubenswrapper[4813]: I0317 09:09:45.144241 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:45 crc kubenswrapper[4813]: I0317 09:09:45.144266 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:45 crc kubenswrapper[4813]: I0317 09:09:45.144308 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 09:09:45 crc kubenswrapper[4813]: I0317 09:09:45.842199 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:45 crc kubenswrapper[4813]: I0317 09:09:45.842206 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:45 crc kubenswrapper[4813]: I0317 09:09:45.843639 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:45 crc kubenswrapper[4813]: I0317 09:09:45.843685 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:45 crc kubenswrapper[4813]: I0317 09:09:45.843697 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:45 crc kubenswrapper[4813]: I0317 09:09:45.843849 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:45 crc kubenswrapper[4813]: I0317 09:09:45.843968 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:45 crc kubenswrapper[4813]: I0317 09:09:45.843991 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:46 crc kubenswrapper[4813]: I0317 09:09:46.616643 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:09:46 crc kubenswrapper[4813]: I0317 09:09:46.845779 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:46 crc kubenswrapper[4813]: I0317 09:09:46.847234 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:46 crc kubenswrapper[4813]: I0317 09:09:46.847308 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:46 crc kubenswrapper[4813]: I0317 09:09:46.847330 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:46 crc kubenswrapper[4813]: I0317 09:09:46.996340 4813 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 09:09:46 crc kubenswrapper[4813]: I0317 09:09:46.996558 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 09:09:48 crc kubenswrapper[4813]: I0317 09:09:48.573283 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:09:48 crc kubenswrapper[4813]: I0317 09:09:48.573494 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:48 crc kubenswrapper[4813]: I0317 09:09:48.575223 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:48 crc kubenswrapper[4813]: I0317 09:09:48.575328 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:48 crc kubenswrapper[4813]: I0317 09:09:48.575351 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:48 crc kubenswrapper[4813]: E0317 09:09:48.819803 4813 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 09:09:49 crc kubenswrapper[4813]: I0317 09:09:49.835253 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 17 09:09:49 crc kubenswrapper[4813]: I0317 09:09:49.835511 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:49 crc kubenswrapper[4813]: I0317 09:09:49.837056 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:49 crc kubenswrapper[4813]: I0317 09:09:49.837174 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:49 crc kubenswrapper[4813]: I0317 09:09:49.837201 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:50 crc kubenswrapper[4813]: I0317 09:09:50.548498 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:09:50 crc kubenswrapper[4813]: I0317 09:09:50.548805 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:50 crc kubenswrapper[4813]: I0317 09:09:50.550337 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:50 crc kubenswrapper[4813]: I0317 09:09:50.550415 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:50 crc kubenswrapper[4813]: I0317 09:09:50.550436 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:50 crc kubenswrapper[4813]: I0317 09:09:50.556687 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:09:50 crc kubenswrapper[4813]: I0317 09:09:50.855520 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:50 crc kubenswrapper[4813]: I0317 09:09:50.857182 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:50 crc kubenswrapper[4813]: I0317 09:09:50.857271 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:50 crc kubenswrapper[4813]: I0317 09:09:50.857363 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:50 crc kubenswrapper[4813]: I0317 09:09:50.864545 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:09:51 crc kubenswrapper[4813]: I0317 09:09:51.858650 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:51 crc kubenswrapper[4813]: I0317 09:09:51.859686 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:51 crc kubenswrapper[4813]: I0317 09:09:51.859733 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:51 crc kubenswrapper[4813]: I0317 09:09:51.859752 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:53 crc kubenswrapper[4813]: W0317 09:09:53.161187 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 17 09:09:53 crc kubenswrapper[4813]: I0317 09:09:53.161340 4813 trace.go:236] Trace[1816585647]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (17-Mar-2026 09:09:43.160) (total time: 10001ms): Mar 17 09:09:53 crc kubenswrapper[4813]: Trace[1816585647]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (09:09:53.161) Mar 17 09:09:53 crc kubenswrapper[4813]: Trace[1816585647]: [10.001284646s] [10.001284646s] END Mar 17 09:09:53 crc kubenswrapper[4813]: E0317 09:09:53.161384 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 17 09:09:53 crc kubenswrapper[4813]: I0317 09:09:53.659250 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Mar 17 09:09:53 crc kubenswrapper[4813]: I0317 09:09:53.762525 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 17 09:09:53 crc kubenswrapper[4813]: I0317 09:09:53.762751 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:53 crc kubenswrapper[4813]: I0317 09:09:53.763792 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:53 crc kubenswrapper[4813]: I0317 09:09:53.763830 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:53 crc kubenswrapper[4813]: I0317 09:09:53.763842 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:53 crc kubenswrapper[4813]: I0317 09:09:53.795689 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 17 09:09:53 crc kubenswrapper[4813]: I0317 09:09:53.863449 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:53 crc kubenswrapper[4813]: I0317 09:09:53.864776 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:53 crc kubenswrapper[4813]: I0317 09:09:53.864835 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:53 crc kubenswrapper[4813]: I0317 09:09:53.864860 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:53 crc kubenswrapper[4813]: I0317 09:09:53.878007 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 17 09:09:54 crc kubenswrapper[4813]: E0317 09:09:54.257960 4813 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:09:54Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189d95d6f0bc735a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.648265562 +0000 UTC m=+0.749069101,LastTimestamp:2026-03-17 09:09:38.648265562 +0000 UTC m=+0.749069101,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:09:54 crc kubenswrapper[4813]: W0317 09:09:54.260360 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:09:54Z is after 2026-02-23T05:33:13Z Mar 17 09:09:54 crc kubenswrapper[4813]: E0317 09:09:54.260475 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:09:54Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 09:09:54 crc kubenswrapper[4813]: W0317 09:09:54.260892 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:09:54Z is after 2026-02-23T05:33:13Z Mar 17 09:09:54 crc kubenswrapper[4813]: E0317 09:09:54.260989 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:09:54Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 09:09:54 crc kubenswrapper[4813]: E0317 09:09:54.261911 4813 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:09:54Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 09:09:54 crc kubenswrapper[4813]: I0317 09:09:54.266883 4813 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 17 09:09:54 crc kubenswrapper[4813]: E0317 09:09:54.266974 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:09:54Z is after 2026-02-23T05:33:13Z" node="crc" Mar 17 09:09:54 crc kubenswrapper[4813]: I0317 09:09:54.266980 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 17 09:09:54 crc kubenswrapper[4813]: W0317 09:09:54.266901 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:09:54Z is after 2026-02-23T05:33:13Z Mar 17 09:09:54 crc kubenswrapper[4813]: E0317 09:09:54.267183 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:09:54Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 09:09:54 crc kubenswrapper[4813]: E0317 09:09:54.270609 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:09:54Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 17 09:09:54 crc kubenswrapper[4813]: I0317 09:09:54.274549 4813 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 17 09:09:54 crc kubenswrapper[4813]: I0317 09:09:54.274627 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 17 09:09:54 crc kubenswrapper[4813]: I0317 09:09:54.661919 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:09:54Z is after 2026-02-23T05:33:13Z Mar 17 09:09:54 crc kubenswrapper[4813]: I0317 09:09:54.867960 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 17 09:09:54 crc kubenswrapper[4813]: I0317 09:09:54.868650 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 17 09:09:54 crc kubenswrapper[4813]: I0317 09:09:54.871145 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5ceeff48801f75c62e01ddcf92c4a850d85cf514063f217b6cfd9c0294545b70" exitCode=255 Mar 17 09:09:54 crc kubenswrapper[4813]: I0317 09:09:54.871259 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"5ceeff48801f75c62e01ddcf92c4a850d85cf514063f217b6cfd9c0294545b70"} Mar 17 09:09:54 crc kubenswrapper[4813]: I0317 09:09:54.871361 4813 scope.go:117] "RemoveContainer" containerID="38fa2834a5940c4d76b73ca7a545601a5344a74c4bd8f5c77f02cf6044d84a0d" Mar 17 09:09:54 crc kubenswrapper[4813]: I0317 09:09:54.871507 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:54 crc kubenswrapper[4813]: I0317 09:09:54.871570 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:54 crc kubenswrapper[4813]: I0317 09:09:54.872824 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:54 crc kubenswrapper[4813]: I0317 09:09:54.872925 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:54 crc kubenswrapper[4813]: I0317 09:09:54.872996 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:54 crc kubenswrapper[4813]: I0317 09:09:54.872877 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:54 crc kubenswrapper[4813]: I0317 09:09:54.873149 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:54 crc kubenswrapper[4813]: I0317 09:09:54.873182 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:54 crc kubenswrapper[4813]: I0317 09:09:54.879159 4813 scope.go:117] "RemoveContainer" containerID="5ceeff48801f75c62e01ddcf92c4a850d85cf514063f217b6cfd9c0294545b70" Mar 17 09:09:54 crc kubenswrapper[4813]: E0317 09:09:54.879497 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:09:55 crc kubenswrapper[4813]: I0317 09:09:55.664117 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:09:55Z is after 2026-02-23T05:33:13Z Mar 17 09:09:55 crc kubenswrapper[4813]: I0317 09:09:55.876046 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 17 09:09:56 crc kubenswrapper[4813]: I0317 09:09:56.624821 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:09:56 crc kubenswrapper[4813]: I0317 09:09:56.624996 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:56 crc kubenswrapper[4813]: I0317 09:09:56.626465 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:56 crc kubenswrapper[4813]: I0317 09:09:56.626516 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:56 crc kubenswrapper[4813]: I0317 09:09:56.626535 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:56 crc kubenswrapper[4813]: I0317 09:09:56.627251 4813 scope.go:117] "RemoveContainer" containerID="5ceeff48801f75c62e01ddcf92c4a850d85cf514063f217b6cfd9c0294545b70" Mar 17 09:09:56 crc kubenswrapper[4813]: E0317 09:09:56.627508 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:09:56 crc kubenswrapper[4813]: I0317 09:09:56.632559 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:09:56 crc kubenswrapper[4813]: I0317 09:09:56.663413 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:09:56Z is after 2026-02-23T05:33:13Z Mar 17 09:09:56 crc kubenswrapper[4813]: I0317 09:09:56.881383 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:09:56 crc kubenswrapper[4813]: I0317 09:09:56.882636 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:09:56 crc kubenswrapper[4813]: I0317 09:09:56.882695 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:09:56 crc kubenswrapper[4813]: I0317 09:09:56.882715 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:09:56 crc kubenswrapper[4813]: I0317 09:09:56.883709 4813 scope.go:117] "RemoveContainer" containerID="5ceeff48801f75c62e01ddcf92c4a850d85cf514063f217b6cfd9c0294545b70" Mar 17 09:09:56 crc kubenswrapper[4813]: E0317 09:09:56.884008 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:09:56 crc kubenswrapper[4813]: I0317 09:09:56.996630 4813 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 09:09:56 crc kubenswrapper[4813]: I0317 09:09:56.996706 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 09:09:57 crc kubenswrapper[4813]: I0317 09:09:57.663020 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:09:57Z is after 2026-02-23T05:33:13Z Mar 17 09:09:58 crc kubenswrapper[4813]: I0317 09:09:58.663104 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:09:58Z is after 2026-02-23T05:33:13Z Mar 17 09:09:58 crc kubenswrapper[4813]: E0317 09:09:58.820014 4813 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 09:09:58 crc kubenswrapper[4813]: W0317 09:09:58.986222 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:09:58Z is after 2026-02-23T05:33:13Z Mar 17 09:09:58 crc kubenswrapper[4813]: E0317 09:09:58.986343 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:09:58Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 09:09:59 crc kubenswrapper[4813]: I0317 09:09:59.661841 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:09:59Z is after 2026-02-23T05:33:13Z Mar 17 09:10:00 crc kubenswrapper[4813]: I0317 09:10:00.664481 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:00Z is after 2026-02-23T05:33:13Z Mar 17 09:10:00 crc kubenswrapper[4813]: I0317 09:10:00.667718 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:00 crc kubenswrapper[4813]: I0317 09:10:00.669397 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:00 crc kubenswrapper[4813]: I0317 09:10:00.669509 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:00 crc kubenswrapper[4813]: I0317 09:10:00.669529 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:00 crc kubenswrapper[4813]: I0317 09:10:00.669574 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 09:10:00 crc kubenswrapper[4813]: E0317 09:10:00.677935 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:00Z is after 2026-02-23T05:33:13Z" node="crc" Mar 17 09:10:00 crc kubenswrapper[4813]: E0317 09:10:00.683061 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:00Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 17 09:10:01 crc kubenswrapper[4813]: W0317 09:10:01.410908 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:01Z is after 2026-02-23T05:33:13Z Mar 17 09:10:01 crc kubenswrapper[4813]: E0317 09:10:01.411023 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:01Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 09:10:01 crc kubenswrapper[4813]: I0317 09:10:01.665235 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:01Z is after 2026-02-23T05:33:13Z Mar 17 09:10:02 crc kubenswrapper[4813]: I0317 09:10:02.083747 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:10:02 crc kubenswrapper[4813]: I0317 09:10:02.084121 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:02 crc kubenswrapper[4813]: I0317 09:10:02.085847 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:02 crc kubenswrapper[4813]: I0317 09:10:02.085897 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:02 crc kubenswrapper[4813]: I0317 09:10:02.085915 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:02 crc kubenswrapper[4813]: I0317 09:10:02.086807 4813 scope.go:117] "RemoveContainer" containerID="5ceeff48801f75c62e01ddcf92c4a850d85cf514063f217b6cfd9c0294545b70" Mar 17 09:10:02 crc kubenswrapper[4813]: E0317 09:10:02.087079 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:10:02 crc kubenswrapper[4813]: I0317 09:10:02.663137 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:02Z is after 2026-02-23T05:33:13Z Mar 17 09:10:02 crc kubenswrapper[4813]: I0317 09:10:02.944437 4813 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 17 09:10:02 crc kubenswrapper[4813]: E0317 09:10:02.949767 4813 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:02Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 09:10:03 crc kubenswrapper[4813]: I0317 09:10:03.663110 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:03Z is after 2026-02-23T05:33:13Z Mar 17 09:10:04 crc kubenswrapper[4813]: E0317 09:10:04.264668 4813 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:04Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189d95d6f0bc735a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.648265562 +0000 UTC m=+0.749069101,LastTimestamp:2026-03-17 09:09:38.648265562 +0000 UTC m=+0.749069101,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:04 crc kubenswrapper[4813]: I0317 09:10:04.663271 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:04Z is after 2026-02-23T05:33:13Z Mar 17 09:10:05 crc kubenswrapper[4813]: W0317 09:10:05.438645 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:05Z is after 2026-02-23T05:33:13Z Mar 17 09:10:05 crc kubenswrapper[4813]: E0317 09:10:05.438784 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:05Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 09:10:05 crc kubenswrapper[4813]: I0317 09:10:05.663308 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:05Z is after 2026-02-23T05:33:13Z Mar 17 09:10:06 crc kubenswrapper[4813]: W0317 09:10:06.540589 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:06Z is after 2026-02-23T05:33:13Z Mar 17 09:10:06 crc kubenswrapper[4813]: E0317 09:10:06.540763 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:06Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 09:10:06 crc kubenswrapper[4813]: I0317 09:10:06.663581 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:06Z is after 2026-02-23T05:33:13Z Mar 17 09:10:06 crc kubenswrapper[4813]: I0317 09:10:06.996983 4813 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 09:10:06 crc kubenswrapper[4813]: I0317 09:10:06.997246 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 09:10:06 crc kubenswrapper[4813]: I0317 09:10:06.997365 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:10:06 crc kubenswrapper[4813]: I0317 09:10:06.997593 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:06 crc kubenswrapper[4813]: I0317 09:10:06.999458 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:06 crc kubenswrapper[4813]: I0317 09:10:06.999509 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:06 crc kubenswrapper[4813]: I0317 09:10:06.999527 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:07 crc kubenswrapper[4813]: I0317 09:10:07.000406 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"6a5b5d0590fefa483bcb5d5a78dfee7d61881fe9e19e5f229d3788d2cc09ab87"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 17 09:10:07 crc kubenswrapper[4813]: I0317 09:10:07.000670 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://6a5b5d0590fefa483bcb5d5a78dfee7d61881fe9e19e5f229d3788d2cc09ab87" gracePeriod=30 Mar 17 09:10:07 crc kubenswrapper[4813]: I0317 09:10:07.662095 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:07Z is after 2026-02-23T05:33:13Z Mar 17 09:10:07 crc kubenswrapper[4813]: I0317 09:10:07.678571 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:07 crc kubenswrapper[4813]: I0317 09:10:07.680671 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:07 crc kubenswrapper[4813]: I0317 09:10:07.680745 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:07 crc kubenswrapper[4813]: I0317 09:10:07.680766 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:07 crc kubenswrapper[4813]: I0317 09:10:07.680799 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 09:10:07 crc kubenswrapper[4813]: E0317 09:10:07.685587 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:07Z is after 2026-02-23T05:33:13Z" node="crc" Mar 17 09:10:07 crc kubenswrapper[4813]: E0317 09:10:07.689218 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:07Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 17 09:10:07 crc kubenswrapper[4813]: W0317 09:10:07.846787 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:07Z is after 2026-02-23T05:33:13Z Mar 17 09:10:07 crc kubenswrapper[4813]: E0317 09:10:07.846892 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:07Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 09:10:07 crc kubenswrapper[4813]: I0317 09:10:07.915782 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 17 09:10:07 crc kubenswrapper[4813]: I0317 09:10:07.916429 4813 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="6a5b5d0590fefa483bcb5d5a78dfee7d61881fe9e19e5f229d3788d2cc09ab87" exitCode=255 Mar 17 09:10:07 crc kubenswrapper[4813]: I0317 09:10:07.916491 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"6a5b5d0590fefa483bcb5d5a78dfee7d61881fe9e19e5f229d3788d2cc09ab87"} Mar 17 09:10:07 crc kubenswrapper[4813]: I0317 09:10:07.916592 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"80403c9745d70069a406ca4771940c9ff6958417787d62595bf005db27d2c1da"} Mar 17 09:10:07 crc kubenswrapper[4813]: I0317 09:10:07.916805 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:07 crc kubenswrapper[4813]: I0317 09:10:07.918138 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:07 crc kubenswrapper[4813]: I0317 09:10:07.918194 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:07 crc kubenswrapper[4813]: I0317 09:10:07.918207 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:08 crc kubenswrapper[4813]: I0317 09:10:08.574006 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:10:08 crc kubenswrapper[4813]: I0317 09:10:08.662902 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:08Z is after 2026-02-23T05:33:13Z Mar 17 09:10:08 crc kubenswrapper[4813]: E0317 09:10:08.820135 4813 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 09:10:08 crc kubenswrapper[4813]: I0317 09:10:08.919817 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:08 crc kubenswrapper[4813]: I0317 09:10:08.921066 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:08 crc kubenswrapper[4813]: I0317 09:10:08.921120 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:08 crc kubenswrapper[4813]: I0317 09:10:08.921137 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:09 crc kubenswrapper[4813]: I0317 09:10:09.663484 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:09Z is after 2026-02-23T05:33:13Z Mar 17 09:10:10 crc kubenswrapper[4813]: I0317 09:10:10.663059 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:10Z is after 2026-02-23T05:33:13Z Mar 17 09:10:11 crc kubenswrapper[4813]: I0317 09:10:11.660924 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:11Z is after 2026-02-23T05:33:13Z Mar 17 09:10:12 crc kubenswrapper[4813]: I0317 09:10:12.659974 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:12Z is after 2026-02-23T05:33:13Z Mar 17 09:10:13 crc kubenswrapper[4813]: I0317 09:10:13.663236 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:13Z is after 2026-02-23T05:33:13Z Mar 17 09:10:13 crc kubenswrapper[4813]: I0317 09:10:13.996143 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:10:13 crc kubenswrapper[4813]: I0317 09:10:13.996487 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:13 crc kubenswrapper[4813]: I0317 09:10:13.997934 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:13 crc kubenswrapper[4813]: I0317 09:10:13.997986 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:13 crc kubenswrapper[4813]: I0317 09:10:13.998000 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:14 crc kubenswrapper[4813]: E0317 09:10:14.271651 4813 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:14Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189d95d6f0bc735a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.648265562 +0000 UTC m=+0.749069101,LastTimestamp:2026-03-17 09:09:38.648265562 +0000 UTC m=+0.749069101,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:14 crc kubenswrapper[4813]: I0317 09:10:14.663820 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:14Z is after 2026-02-23T05:33:13Z Mar 17 09:10:14 crc kubenswrapper[4813]: I0317 09:10:14.686483 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:14 crc kubenswrapper[4813]: I0317 09:10:14.688061 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:14 crc kubenswrapper[4813]: I0317 09:10:14.688139 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:14 crc kubenswrapper[4813]: I0317 09:10:14.688160 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:14 crc kubenswrapper[4813]: I0317 09:10:14.688197 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 09:10:14 crc kubenswrapper[4813]: E0317 09:10:14.692928 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:14Z is after 2026-02-23T05:33:13Z" node="crc" Mar 17 09:10:14 crc kubenswrapper[4813]: E0317 09:10:14.696033 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:14Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 17 09:10:15 crc kubenswrapper[4813]: I0317 09:10:15.663382 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:15Z is after 2026-02-23T05:33:13Z Mar 17 09:10:16 crc kubenswrapper[4813]: I0317 09:10:16.661438 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:16Z is after 2026-02-23T05:33:13Z Mar 17 09:10:16 crc kubenswrapper[4813]: I0317 09:10:16.730260 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:16 crc kubenswrapper[4813]: I0317 09:10:16.732365 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:16 crc kubenswrapper[4813]: I0317 09:10:16.732430 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:16 crc kubenswrapper[4813]: I0317 09:10:16.732448 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:16 crc kubenswrapper[4813]: I0317 09:10:16.733306 4813 scope.go:117] "RemoveContainer" containerID="5ceeff48801f75c62e01ddcf92c4a850d85cf514063f217b6cfd9c0294545b70" Mar 17 09:10:16 crc kubenswrapper[4813]: W0317 09:10:16.808247 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:16Z is after 2026-02-23T05:33:13Z Mar 17 09:10:16 crc kubenswrapper[4813]: E0317 09:10:16.808327 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:16Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 09:10:16 crc kubenswrapper[4813]: I0317 09:10:16.996488 4813 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 09:10:16 crc kubenswrapper[4813]: I0317 09:10:16.996558 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 09:10:17 crc kubenswrapper[4813]: I0317 09:10:17.663130 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:17Z is after 2026-02-23T05:33:13Z Mar 17 09:10:17 crc kubenswrapper[4813]: I0317 09:10:17.949515 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 17 09:10:17 crc kubenswrapper[4813]: I0317 09:10:17.950202 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 17 09:10:17 crc kubenswrapper[4813]: I0317 09:10:17.952425 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="056a72ec3942647bc662c27e16e4d06a6925fe85f46704866499d192b596af48" exitCode=255 Mar 17 09:10:17 crc kubenswrapper[4813]: I0317 09:10:17.952485 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"056a72ec3942647bc662c27e16e4d06a6925fe85f46704866499d192b596af48"} Mar 17 09:10:17 crc kubenswrapper[4813]: I0317 09:10:17.952537 4813 scope.go:117] "RemoveContainer" containerID="5ceeff48801f75c62e01ddcf92c4a850d85cf514063f217b6cfd9c0294545b70" Mar 17 09:10:17 crc kubenswrapper[4813]: I0317 09:10:17.952804 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:17 crc kubenswrapper[4813]: I0317 09:10:17.954158 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:17 crc kubenswrapper[4813]: I0317 09:10:17.954196 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:17 crc kubenswrapper[4813]: I0317 09:10:17.954215 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:17 crc kubenswrapper[4813]: I0317 09:10:17.955231 4813 scope.go:117] "RemoveContainer" containerID="056a72ec3942647bc662c27e16e4d06a6925fe85f46704866499d192b596af48" Mar 17 09:10:17 crc kubenswrapper[4813]: E0317 09:10:17.955539 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:10:18 crc kubenswrapper[4813]: I0317 09:10:18.663009 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:18Z is after 2026-02-23T05:33:13Z Mar 17 09:10:18 crc kubenswrapper[4813]: E0317 09:10:18.820524 4813 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 09:10:18 crc kubenswrapper[4813]: I0317 09:10:18.957702 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 17 09:10:19 crc kubenswrapper[4813]: I0317 09:10:19.565258 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:10:19 crc kubenswrapper[4813]: I0317 09:10:19.565444 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:19 crc kubenswrapper[4813]: I0317 09:10:19.566697 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:19 crc kubenswrapper[4813]: I0317 09:10:19.566741 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:19 crc kubenswrapper[4813]: I0317 09:10:19.566757 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:19 crc kubenswrapper[4813]: I0317 09:10:19.567355 4813 scope.go:117] "RemoveContainer" containerID="056a72ec3942647bc662c27e16e4d06a6925fe85f46704866499d192b596af48" Mar 17 09:10:19 crc kubenswrapper[4813]: E0317 09:10:19.567579 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:10:19 crc kubenswrapper[4813]: I0317 09:10:19.661392 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:19Z is after 2026-02-23T05:33:13Z Mar 17 09:10:20 crc kubenswrapper[4813]: I0317 09:10:20.363594 4813 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 17 09:10:20 crc kubenswrapper[4813]: E0317 09:10:20.368009 4813 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:20Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 09:10:20 crc kubenswrapper[4813]: E0317 09:10:20.369244 4813 certificate_manager.go:440] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Reached backoff limit, still unable to rotate certs: timed out waiting for the condition" logger="UnhandledError" Mar 17 09:10:20 crc kubenswrapper[4813]: I0317 09:10:20.662805 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:20Z is after 2026-02-23T05:33:13Z Mar 17 09:10:21 crc kubenswrapper[4813]: I0317 09:10:21.676776 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:21Z is after 2026-02-23T05:33:13Z Mar 17 09:10:21 crc kubenswrapper[4813]: I0317 09:10:21.694143 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:21 crc kubenswrapper[4813]: I0317 09:10:21.695950 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:21 crc kubenswrapper[4813]: I0317 09:10:21.695998 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:21 crc kubenswrapper[4813]: I0317 09:10:21.696015 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:21 crc kubenswrapper[4813]: I0317 09:10:21.696047 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 09:10:21 crc kubenswrapper[4813]: E0317 09:10:21.701111 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:21Z is after 2026-02-23T05:33:13Z" interval="7s" Mar 17 09:10:21 crc kubenswrapper[4813]: E0317 09:10:21.705505 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:21Z is after 2026-02-23T05:33:13Z" node="crc" Mar 17 09:10:22 crc kubenswrapper[4813]: W0317 09:10:22.074387 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:22Z is after 2026-02-23T05:33:13Z Mar 17 09:10:22 crc kubenswrapper[4813]: E0317 09:10:22.074492 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:22Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 09:10:22 crc kubenswrapper[4813]: I0317 09:10:22.084058 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:10:22 crc kubenswrapper[4813]: I0317 09:10:22.084352 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:22 crc kubenswrapper[4813]: I0317 09:10:22.086015 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:22 crc kubenswrapper[4813]: I0317 09:10:22.086072 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:22 crc kubenswrapper[4813]: I0317 09:10:22.086089 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:22 crc kubenswrapper[4813]: I0317 09:10:22.086899 4813 scope.go:117] "RemoveContainer" containerID="056a72ec3942647bc662c27e16e4d06a6925fe85f46704866499d192b596af48" Mar 17 09:10:22 crc kubenswrapper[4813]: E0317 09:10:22.087168 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:10:22 crc kubenswrapper[4813]: I0317 09:10:22.663020 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:22Z is after 2026-02-23T05:33:13Z Mar 17 09:10:23 crc kubenswrapper[4813]: I0317 09:10:23.662976 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:23Z is after 2026-02-23T05:33:13Z Mar 17 09:10:24 crc kubenswrapper[4813]: E0317 09:10:24.277164 4813 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:24Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189d95d6f0bc735a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.648265562 +0000 UTC m=+0.749069101,LastTimestamp:2026-03-17 09:09:38.648265562 +0000 UTC m=+0.749069101,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:24 crc kubenswrapper[4813]: I0317 09:10:24.661550 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:24Z is after 2026-02-23T05:33:13Z Mar 17 09:10:25 crc kubenswrapper[4813]: W0317 09:10:25.184124 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:25Z is after 2026-02-23T05:33:13Z Mar 17 09:10:25 crc kubenswrapper[4813]: E0317 09:10:25.184239 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:25Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 17 09:10:25 crc kubenswrapper[4813]: I0317 09:10:25.661374 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:25Z is after 2026-02-23T05:33:13Z Mar 17 09:10:26 crc kubenswrapper[4813]: I0317 09:10:26.664110 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:26 crc kubenswrapper[4813]: I0317 09:10:26.997183 4813 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 09:10:26 crc kubenswrapper[4813]: I0317 09:10:26.997286 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 09:10:27 crc kubenswrapper[4813]: I0317 09:10:27.663545 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:28 crc kubenswrapper[4813]: I0317 09:10:28.662457 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:28 crc kubenswrapper[4813]: I0317 09:10:28.705735 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:28 crc kubenswrapper[4813]: I0317 09:10:28.707995 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:28 crc kubenswrapper[4813]: I0317 09:10:28.708091 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:28 crc kubenswrapper[4813]: I0317 09:10:28.708120 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:28 crc kubenswrapper[4813]: I0317 09:10:28.708172 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 09:10:28 crc kubenswrapper[4813]: E0317 09:10:28.709230 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 17 09:10:28 crc kubenswrapper[4813]: E0317 09:10:28.715872 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 17 09:10:28 crc kubenswrapper[4813]: E0317 09:10:28.821642 4813 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 09:10:29 crc kubenswrapper[4813]: I0317 09:10:29.662269 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:30 crc kubenswrapper[4813]: W0317 09:10:30.663551 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:30 crc kubenswrapper[4813]: I0317 09:10:30.663639 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:30 crc kubenswrapper[4813]: E0317 09:10:30.663638 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 17 09:10:31 crc kubenswrapper[4813]: I0317 09:10:31.665455 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:32 crc kubenswrapper[4813]: I0317 09:10:32.664868 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:32 crc kubenswrapper[4813]: I0317 09:10:32.730676 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:32 crc kubenswrapper[4813]: I0317 09:10:32.732295 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:32 crc kubenswrapper[4813]: I0317 09:10:32.732522 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:32 crc kubenswrapper[4813]: I0317 09:10:32.732764 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:32 crc kubenswrapper[4813]: I0317 09:10:32.734451 4813 scope.go:117] "RemoveContainer" containerID="056a72ec3942647bc662c27e16e4d06a6925fe85f46704866499d192b596af48" Mar 17 09:10:32 crc kubenswrapper[4813]: E0317 09:10:32.735315 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:10:33 crc kubenswrapper[4813]: I0317 09:10:33.665554 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:33 crc kubenswrapper[4813]: I0317 09:10:33.730878 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 17 09:10:33 crc kubenswrapper[4813]: I0317 09:10:33.731094 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:33 crc kubenswrapper[4813]: I0317 09:10:33.732582 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:33 crc kubenswrapper[4813]: I0317 09:10:33.732715 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:33 crc kubenswrapper[4813]: I0317 09:10:33.732746 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.285834 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f0bc735a default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.648265562 +0000 UTC m=+0.749069101,LastTimestamp:2026-03-17 09:09:38.648265562 +0000 UTC m=+0.749069101,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.292777 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f522f847 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722093127 +0000 UTC m=+0.822896636,LastTimestamp:2026-03-17 09:09:38.722093127 +0000 UTC m=+0.822896636,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.299347 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f5234b86 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722114438 +0000 UTC m=+0.822917947,LastTimestamp:2026-03-17 09:09:38.722114438 +0000 UTC m=+0.822917947,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.306669 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f52376ec default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722125548 +0000 UTC m=+0.822929057,LastTimestamp:2026-03-17 09:09:38.722125548 +0000 UTC m=+0.822929057,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.313958 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6fac6d191 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.816717201 +0000 UTC m=+0.917520740,LastTimestamp:2026-03-17 09:09:38.816717201 +0000 UTC m=+0.917520740,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.321229 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d95d6f522f847\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f522f847 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722093127 +0000 UTC m=+0.822896636,LastTimestamp:2026-03-17 09:09:38.831197706 +0000 UTC m=+0.932001215,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.327903 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d95d6f5234b86\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f5234b86 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722114438 +0000 UTC m=+0.822917947,LastTimestamp:2026-03-17 09:09:38.831228446 +0000 UTC m=+0.932031955,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.335021 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d95d6f52376ec\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f52376ec default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722125548 +0000 UTC m=+0.822929057,LastTimestamp:2026-03-17 09:09:38.831239557 +0000 UTC m=+0.932043066,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.343305 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d95d6f522f847\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f522f847 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722093127 +0000 UTC m=+0.822896636,LastTimestamp:2026-03-17 09:09:38.832087543 +0000 UTC m=+0.932891042,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.351402 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d95d6f5234b86\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f5234b86 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722114438 +0000 UTC m=+0.822917947,LastTimestamp:2026-03-17 09:09:38.832107804 +0000 UTC m=+0.932911303,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.359022 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d95d6f52376ec\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f52376ec default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722125548 +0000 UTC m=+0.822929057,LastTimestamp:2026-03-17 09:09:38.832115864 +0000 UTC m=+0.932919363,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.364015 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d95d6f522f847\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f522f847 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722093127 +0000 UTC m=+0.822896636,LastTimestamp:2026-03-17 09:09:38.832522847 +0000 UTC m=+0.933326356,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.369814 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d95d6f5234b86\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f5234b86 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722114438 +0000 UTC m=+0.822917947,LastTimestamp:2026-03-17 09:09:38.832537267 +0000 UTC m=+0.933340776,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.376491 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d95d6f52376ec\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f52376ec default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722125548 +0000 UTC m=+0.822929057,LastTimestamp:2026-03-17 09:09:38.83263545 +0000 UTC m=+0.933438949,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.383972 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d95d6f522f847\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f522f847 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722093127 +0000 UTC m=+0.822896636,LastTimestamp:2026-03-17 09:09:38.832887348 +0000 UTC m=+0.933690847,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.390682 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d95d6f5234b86\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f5234b86 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722114438 +0000 UTC m=+0.822917947,LastTimestamp:2026-03-17 09:09:38.832897068 +0000 UTC m=+0.933700567,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.397353 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d95d6f52376ec\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f52376ec default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722125548 +0000 UTC m=+0.822929057,LastTimestamp:2026-03-17 09:09:38.832904738 +0000 UTC m=+0.933708237,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.404550 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d95d6f522f847\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f522f847 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722093127 +0000 UTC m=+0.822896636,LastTimestamp:2026-03-17 09:09:38.833526517 +0000 UTC m=+0.934330026,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.410696 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d95d6f5234b86\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f5234b86 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722114438 +0000 UTC m=+0.822917947,LastTimestamp:2026-03-17 09:09:38.833547917 +0000 UTC m=+0.934351426,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.415407 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d95d6f52376ec\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f52376ec default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722125548 +0000 UTC m=+0.822929057,LastTimestamp:2026-03-17 09:09:38.833559248 +0000 UTC m=+0.934362757,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.421511 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d95d6f522f847\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f522f847 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722093127 +0000 UTC m=+0.822896636,LastTimestamp:2026-03-17 09:09:38.834859898 +0000 UTC m=+0.935663437,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.427975 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d95d6f5234b86\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f5234b86 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722114438 +0000 UTC m=+0.822917947,LastTimestamp:2026-03-17 09:09:38.834880259 +0000 UTC m=+0.935683788,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.435266 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d95d6f52376ec\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f52376ec default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722125548 +0000 UTC m=+0.822929057,LastTimestamp:2026-03-17 09:09:38.834895599 +0000 UTC m=+0.935699128,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.441137 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d95d6f522f847\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f522f847 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722093127 +0000 UTC m=+0.822896636,LastTimestamp:2026-03-17 09:09:38.835041294 +0000 UTC m=+0.935844813,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.446756 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189d95d6f5234b86\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189d95d6f5234b86 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:38.722114438 +0000 UTC m=+0.822917947,LastTimestamp:2026-03-17 09:09:38.835055514 +0000 UTC m=+0.935859033,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.451683 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d95d715d3ed7d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:39.270561149 +0000 UTC m=+1.371364658,LastTimestamp:2026-03-17 09:09:39.270561149 +0000 UTC m=+1.371364658,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.458315 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d95d715db2a5c openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:39.271035484 +0000 UTC m=+1.371839023,LastTimestamp:2026-03-17 09:09:39.271035484 +0000 UTC m=+1.371839023,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.461922 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189d95d715e1535c openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:39.271439196 +0000 UTC m=+1.372242705,LastTimestamp:2026-03-17 09:09:39.271439196 +0000 UTC m=+1.372242705,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.465749 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d7170eb3ec openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:39.291190252 +0000 UTC m=+1.391993761,LastTimestamp:2026-03-17 09:09:39.291190252 +0000 UTC m=+1.391993761,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.469885 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95d717d28aa0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:39.304024736 +0000 UTC m=+1.404828245,LastTimestamp:2026-03-17 09:09:39.304024736 +0000 UTC m=+1.404828245,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.476644 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d95d73a22541a openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:39.879679002 +0000 UTC m=+1.980482511,LastTimestamp:2026-03-17 09:09:39.879679002 +0000 UTC m=+1.980482511,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.483543 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d73a297f1f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:39.880148767 +0000 UTC m=+1.980952276,LastTimestamp:2026-03-17 09:09:39.880148767 +0000 UTC m=+1.980952276,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.488107 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189d95d73a2bb17b openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:39.880292731 +0000 UTC m=+1.981096240,LastTimestamp:2026-03-17 09:09:39.880292731 +0000 UTC m=+1.981096240,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.492641 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d95d73a2c4d8a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:39.880332682 +0000 UTC m=+1.981136191,LastTimestamp:2026-03-17 09:09:39.880332682 +0000 UTC m=+1.981136191,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.497455 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95d73a60a625 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:39.883763237 +0000 UTC m=+1.984566746,LastTimestamp:2026-03-17 09:09:39.883763237 +0000 UTC m=+1.984566746,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.502105 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d95d73b22d048 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:39.896488008 +0000 UTC m=+1.997291517,LastTimestamp:2026-03-17 09:09:39.896488008 +0000 UTC m=+1.997291517,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.509007 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d95d73b3e8564 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:39.898303844 +0000 UTC m=+1.999107353,LastTimestamp:2026-03-17 09:09:39.898303844 +0000 UTC m=+1.999107353,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.515660 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d95d73b448a2e openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:39.898698286 +0000 UTC m=+1.999501825,LastTimestamp:2026-03-17 09:09:39.898698286 +0000 UTC m=+1.999501825,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.523274 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189d95d73b567dc9 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:39.899874761 +0000 UTC m=+2.000678270,LastTimestamp:2026-03-17 09:09:39.899874761 +0000 UTC m=+2.000678270,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.529953 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95d73b791010 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:39.902140432 +0000 UTC m=+2.002943941,LastTimestamp:2026-03-17 09:09:39.902140432 +0000 UTC m=+2.002943941,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.536711 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d73b876d3e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:39.90308179 +0000 UTC m=+2.003885299,LastTimestamp:2026-03-17 09:09:39.90308179 +0000 UTC m=+2.003885299,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.543437 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d95d74f01cda9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:40.229868969 +0000 UTC m=+2.330672508,LastTimestamp:2026-03-17 09:09:40.229868969 +0000 UTC m=+2.330672508,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.550543 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d95d74fd89813 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:40.243945491 +0000 UTC m=+2.344748990,LastTimestamp:2026-03-17 09:09:40.243945491 +0000 UTC m=+2.344748990,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.557369 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d95d74ff41660 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:40.245747296 +0000 UTC m=+2.346550835,LastTimestamp:2026-03-17 09:09:40.245747296 +0000 UTC m=+2.346550835,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.565050 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d95d75c140292 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:40.44916597 +0000 UTC m=+2.549969509,LastTimestamp:2026-03-17 09:09:40.44916597 +0000 UTC m=+2.549969509,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.571468 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d95d75e3e483c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:40.485490748 +0000 UTC m=+2.586294287,LastTimestamp:2026-03-17 09:09:40.485490748 +0000 UTC m=+2.586294287,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.578403 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d95d75e61d981 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:40.487821697 +0000 UTC m=+2.588625236,LastTimestamp:2026-03-17 09:09:40.487821697 +0000 UTC m=+2.588625236,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.584830 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d95d76d1c4a66 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:40.734921318 +0000 UTC m=+2.835724817,LastTimestamp:2026-03-17 09:09:40.734921318 +0000 UTC m=+2.835724817,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.592423 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d95d76dcdb3d8 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:40.746548184 +0000 UTC m=+2.847351693,LastTimestamp:2026-03-17 09:09:40.746548184 +0000 UTC m=+2.847351693,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.599952 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d95d76e7e2908 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:40.75811252 +0000 UTC m=+2.858916029,LastTimestamp:2026-03-17 09:09:40.75811252 +0000 UTC m=+2.858916029,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.608696 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95d76f237dc8 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:40.768947656 +0000 UTC m=+2.869751175,LastTimestamp:2026-03-17 09:09:40.768947656 +0000 UTC m=+2.869751175,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.617225 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d76f5d9722 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:40.772755234 +0000 UTC m=+2.873558733,LastTimestamp:2026-03-17 09:09:40.772755234 +0000 UTC m=+2.873558733,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.624304 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189d95d76f70a0d4 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:40.7740029 +0000 UTC m=+2.874806409,LastTimestamp:2026-03-17 09:09:40.7740029 +0000 UTC m=+2.874806409,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.631162 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d95d77bbaad2c openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:40.980182316 +0000 UTC m=+3.080985815,LastTimestamp:2026-03-17 09:09:40.980182316 +0000 UTC m=+3.080985815,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.637370 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95d77c879916 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:40.993612054 +0000 UTC m=+3.094415553,LastTimestamp:2026-03-17 09:09:40.993612054 +0000 UTC m=+3.094415553,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.642035 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d95d77ce4ecb8 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:40.999728312 +0000 UTC m=+3.100531811,LastTimestamp:2026-03-17 09:09:40.999728312 +0000 UTC m=+3.100531811,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.646351 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d95d77cf652cb openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.000868555 +0000 UTC m=+3.101672064,LastTimestamp:2026-03-17 09:09:41.000868555 +0000 UTC m=+3.101672064,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.649997 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189d95d77d0934ff openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.002106111 +0000 UTC m=+3.102909610,LastTimestamp:2026-03-17 09:09:41.002106111 +0000 UTC m=+3.102909610,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.653275 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95d77d59912c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.007372588 +0000 UTC m=+3.108176087,LastTimestamp:2026-03-17 09:09:41.007372588 +0000 UTC m=+3.108176087,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.656832 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95d77d77b052 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.009346642 +0000 UTC m=+3.110150141,LastTimestamp:2026-03-17 09:09:41.009346642 +0000 UTC m=+3.110150141,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: I0317 09:10:34.660871 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.661150 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d77e2215bb openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.020513723 +0000 UTC m=+3.121317222,LastTimestamp:2026-03-17 09:09:41.020513723 +0000 UTC m=+3.121317222,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.665805 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189d95d77ec0688d openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.030889613 +0000 UTC m=+3.131693122,LastTimestamp:2026-03-17 09:09:41.030889613 +0000 UTC m=+3.131693122,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.669678 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d77fa383e1 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.045773281 +0000 UTC m=+3.146576800,LastTimestamp:2026-03-17 09:09:41.045773281 +0000 UTC m=+3.146576800,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.670514 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d95d788442ab1 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.190519473 +0000 UTC m=+3.291322992,LastTimestamp:2026-03-17 09:09:41.190519473 +0000 UTC m=+3.291322992,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.673721 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95d788729b01 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.193562881 +0000 UTC m=+3.294366380,LastTimestamp:2026-03-17 09:09:41.193562881 +0000 UTC m=+3.294366380,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.677754 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d95d7891f7eb4 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.204893364 +0000 UTC m=+3.305696883,LastTimestamp:2026-03-17 09:09:41.204893364 +0000 UTC m=+3.305696883,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.681042 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d95d789309892 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.206014098 +0000 UTC m=+3.306817597,LastTimestamp:2026-03-17 09:09:41.206014098 +0000 UTC m=+3.306817597,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.685297 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95d789ac20f4 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.21410994 +0000 UTC m=+3.314913439,LastTimestamp:2026-03-17 09:09:41.21410994 +0000 UTC m=+3.314913439,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.688627 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95d789b6a3c5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.214798789 +0000 UTC m=+3.315602298,LastTimestamp:2026-03-17 09:09:41.214798789 +0000 UTC m=+3.315602298,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.692180 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d95d797fff406 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.454484486 +0000 UTC m=+3.555287995,LastTimestamp:2026-03-17 09:09:41.454484486 +0000 UTC m=+3.555287995,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.699940 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95d79873a167 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.462065511 +0000 UTC m=+3.562869000,LastTimestamp:2026-03-17 09:09:41.462065511 +0000 UTC m=+3.562869000,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.703494 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189d95d7996c913d openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.478379837 +0000 UTC m=+3.579183336,LastTimestamp:2026-03-17 09:09:41.478379837 +0000 UTC m=+3.579183336,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.706490 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95d79bbaa8f0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.517052144 +0000 UTC m=+3.617855683,LastTimestamp:2026-03-17 09:09:41.517052144 +0000 UTC m=+3.617855683,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.709473 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95d79bec99bf openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.520325055 +0000 UTC m=+3.621128554,LastTimestamp:2026-03-17 09:09:41.520325055 +0000 UTC m=+3.621128554,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.712402 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95d7aa7b17c1 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.764544449 +0000 UTC m=+3.865347988,LastTimestamp:2026-03-17 09:09:41.764544449 +0000 UTC m=+3.865347988,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.716017 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d7ab7c61d1 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.781406161 +0000 UTC m=+3.882209680,LastTimestamp:2026-03-17 09:09:41.781406161 +0000 UTC m=+3.882209680,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.719114 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95d7ab80128e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.781648014 +0000 UTC m=+3.882451503,LastTimestamp:2026-03-17 09:09:41.781648014 +0000 UTC m=+3.882451503,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.722291 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95d7ab93daf5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.782944501 +0000 UTC m=+3.883748040,LastTimestamp:2026-03-17 09:09:41.782944501 +0000 UTC m=+3.883748040,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.725913 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95d7b7562350 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.980226384 +0000 UTC m=+4.081029883,LastTimestamp:2026-03-17 09:09:41.980226384 +0000 UTC m=+4.081029883,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.728914 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95d7b80f3bb3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.992356787 +0000 UTC m=+4.093160286,LastTimestamp:2026-03-17 09:09:41.992356787 +0000 UTC m=+4.093160286,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.731745 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d7b8348585 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.994800517 +0000 UTC m=+4.095604016,LastTimestamp:2026-03-17 09:09:41.994800517 +0000 UTC m=+4.095604016,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.734828 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d7b985deec openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:42.016909036 +0000 UTC m=+4.117712535,LastTimestamp:2026-03-17 09:09:42.016909036 +0000 UTC m=+4.117712535,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.738473 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189d95d7ab93daf5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95d7ab93daf5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.782944501 +0000 UTC m=+3.883748040,LastTimestamp:2026-03-17 09:09:42.80054608 +0000 UTC m=+4.901349619,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.743071 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d7e89e4c92 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:42.807039122 +0000 UTC m=+4.907842661,LastTimestamp:2026-03-17 09:09:42.807039122 +0000 UTC m=+4.907842661,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.746156 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189d95d7b7562350\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95d7b7562350 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.980226384 +0000 UTC m=+4.081029883,LastTimestamp:2026-03-17 09:09:43.022784808 +0000 UTC m=+5.123588317,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.749248 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d7f57f3980 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:43.023106432 +0000 UTC m=+5.123909941,LastTimestamp:2026-03-17 09:09:43.023106432 +0000 UTC m=+5.123909941,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.752325 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189d95d7b80f3bb3\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95d7b80f3bb3 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:41.992356787 +0000 UTC m=+4.093160286,LastTimestamp:2026-03-17 09:09:43.033166738 +0000 UTC m=+5.133970247,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.755417 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d7f65006f2 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:43.036790514 +0000 UTC m=+5.137594053,LastTimestamp:2026-03-17 09:09:43.036790514 +0000 UTC m=+5.137594053,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.759708 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d7f660e4a9 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:43.037895849 +0000 UTC m=+5.138699358,LastTimestamp:2026-03-17 09:09:43.037895849 +0000 UTC m=+5.138699358,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.763071 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d8048baf25 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:43.275581221 +0000 UTC m=+5.376384770,LastTimestamp:2026-03-17 09:09:43.275581221 +0000 UTC m=+5.376384770,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.767753 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d80576811a openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:43.290970394 +0000 UTC m=+5.391773893,LastTimestamp:2026-03-17 09:09:43.290970394 +0000 UTC m=+5.391773893,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.770768 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d8058908cc openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:43.29218478 +0000 UTC m=+5.392988289,LastTimestamp:2026-03-17 09:09:43.29218478 +0000 UTC m=+5.392988289,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.774959 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d8133ca8f0 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:43.522060528 +0000 UTC m=+5.622864027,LastTimestamp:2026-03-17 09:09:43.522060528 +0000 UTC m=+5.622864027,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.778793 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d813f4fffe openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:43.534141438 +0000 UTC m=+5.634944937,LastTimestamp:2026-03-17 09:09:43.534141438 +0000 UTC m=+5.634944937,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.782508 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d814048e97 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:43.535160983 +0000 UTC m=+5.635964482,LastTimestamp:2026-03-17 09:09:43.535160983 +0000 UTC m=+5.635964482,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.786263 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d820627887 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:43.742642311 +0000 UTC m=+5.843445810,LastTimestamp:2026-03-17 09:09:43.742642311 +0000 UTC m=+5.843445810,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.790169 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d8219512e5 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:43.762735845 +0000 UTC m=+5.863539344,LastTimestamp:2026-03-17 09:09:43.762735845 +0000 UTC m=+5.863539344,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.794061 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d821c4a259 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:43.765852761 +0000 UTC m=+5.866656300,LastTimestamp:2026-03-17 09:09:43.765852761 +0000 UTC m=+5.866656300,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.798280 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d82f51b9c7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:43.993203143 +0000 UTC m=+6.094006662,LastTimestamp:2026-03-17 09:09:43.993203143 +0000 UTC m=+6.094006662,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.802341 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189d95d8300dc699 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:44.005527193 +0000 UTC m=+6.106330732,LastTimestamp:2026-03-17 09:09:44.005527193 +0000 UTC m=+6.106330732,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.807163 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 17 09:10:34 crc kubenswrapper[4813]: &Event{ObjectMeta:{kube-controller-manager-crc.189d95d8e254aa38 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 17 09:10:34 crc kubenswrapper[4813]: body: Mar 17 09:10:34 crc kubenswrapper[4813]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:46.996517432 +0000 UTC m=+9.097320971,LastTimestamp:2026-03-17 09:09:46.996517432 +0000 UTC m=+9.097320971,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 09:10:34 crc kubenswrapper[4813]: > Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.810806 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d95d8e256a3fa openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:46.996646906 +0000 UTC m=+9.097450435,LastTimestamp:2026-03-17 09:09:46.996646906 +0000 UTC m=+9.097450435,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.816658 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 17 09:10:34 crc kubenswrapper[4813]: &Event{ObjectMeta:{kube-apiserver-crc.189d95da93aec1c6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 17 09:10:34 crc kubenswrapper[4813]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 17 09:10:34 crc kubenswrapper[4813]: Mar 17 09:10:34 crc kubenswrapper[4813]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:54.26695623 +0000 UTC m=+16.367759769,LastTimestamp:2026-03-17 09:09:54.26695623 +0000 UTC m=+16.367759769,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 09:10:34 crc kubenswrapper[4813]: > Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.820739 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95da93afc780 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:54.267023232 +0000 UTC m=+16.367826771,LastTimestamp:2026-03-17 09:09:54.267023232 +0000 UTC m=+16.367826771,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.825025 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189d95da93aec1c6\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 17 09:10:34 crc kubenswrapper[4813]: &Event{ObjectMeta:{kube-apiserver-crc.189d95da93aec1c6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 17 09:10:34 crc kubenswrapper[4813]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 17 09:10:34 crc kubenswrapper[4813]: Mar 17 09:10:34 crc kubenswrapper[4813]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:54.26695623 +0000 UTC m=+16.367759769,LastTimestamp:2026-03-17 09:09:54.274588489 +0000 UTC m=+16.375391988,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 09:10:34 crc kubenswrapper[4813]: > Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.828958 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189d95da93afc780\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189d95da93afc780 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:54.267023232 +0000 UTC m=+16.367826771,LastTimestamp:2026-03-17 09:09:54.274650321 +0000 UTC m=+16.375453820,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.834388 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 17 09:10:34 crc kubenswrapper[4813]: &Event{ObjectMeta:{kube-controller-manager-crc.189d95db3663055a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 17 09:10:34 crc kubenswrapper[4813]: body: Mar 17 09:10:34 crc kubenswrapper[4813]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:56.996679002 +0000 UTC m=+19.097482491,LastTimestamp:2026-03-17 09:09:56.996679002 +0000 UTC m=+19.097482491,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 09:10:34 crc kubenswrapper[4813]: > Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.838180 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d95db3663d2ab openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:56.996731563 +0000 UTC m=+19.097535062,LastTimestamp:2026-03-17 09:09:56.996731563 +0000 UTC m=+19.097535062,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.842671 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189d95db3663055a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 17 09:10:34 crc kubenswrapper[4813]: &Event{ObjectMeta:{kube-controller-manager-crc.189d95db3663055a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 17 09:10:34 crc kubenswrapper[4813]: body: Mar 17 09:10:34 crc kubenswrapper[4813]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:56.996679002 +0000 UTC m=+19.097482491,LastTimestamp:2026-03-17 09:10:06.997147267 +0000 UTC m=+29.097950866,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 09:10:34 crc kubenswrapper[4813]: > Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.846623 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189d95db3663d2ab\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d95db3663d2ab openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:56.996731563 +0000 UTC m=+19.097535062,LastTimestamp:2026-03-17 09:10:06.997308213 +0000 UTC m=+29.098111752,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.851133 4813 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d95dd8aab6f78 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:10:07.000645496 +0000 UTC m=+29.101449025,LastTimestamp:2026-03-17 09:10:07.000645496 +0000 UTC m=+29.101449025,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.855212 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189d95d73b3e8564\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d95d73b3e8564 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:39.898303844 +0000 UTC m=+1.999107353,LastTimestamp:2026-03-17 09:10:07.126958923 +0000 UTC m=+29.227762422,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.858441 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189d95d74f01cda9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d95d74f01cda9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:40.229868969 +0000 UTC m=+2.330672508,LastTimestamp:2026-03-17 09:10:07.321133938 +0000 UTC m=+29.421937447,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.862774 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189d95d74fd89813\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d95d74fd89813 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:40.243945491 +0000 UTC m=+2.344748990,LastTimestamp:2026-03-17 09:10:07.329856865 +0000 UTC m=+29.430660374,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.867464 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189d95db3663055a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 17 09:10:34 crc kubenswrapper[4813]: &Event{ObjectMeta:{kube-controller-manager-crc.189d95db3663055a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 17 09:10:34 crc kubenswrapper[4813]: body: Mar 17 09:10:34 crc kubenswrapper[4813]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:56.996679002 +0000 UTC m=+19.097482491,LastTimestamp:2026-03-17 09:10:16.996539118 +0000 UTC m=+39.097342657,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 09:10:34 crc kubenswrapper[4813]: > Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.870843 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189d95db3663d2ab\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189d95db3663d2ab openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:56.996731563 +0000 UTC m=+19.097535062,LastTimestamp:2026-03-17 09:10:16.996587929 +0000 UTC m=+39.097391468,Count:3,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:10:34 crc kubenswrapper[4813]: E0317 09:10:34.874744 4813 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189d95db3663055a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 17 09:10:34 crc kubenswrapper[4813]: &Event{ObjectMeta:{kube-controller-manager-crc.189d95db3663055a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 17 09:10:34 crc kubenswrapper[4813]: body: Mar 17 09:10:34 crc kubenswrapper[4813]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:09:56.996679002 +0000 UTC m=+19.097482491,LastTimestamp:2026-03-17 09:10:26.997258643 +0000 UTC m=+49.098062182,Count:4,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 17 09:10:34 crc kubenswrapper[4813]: > Mar 17 09:10:35 crc kubenswrapper[4813]: I0317 09:10:35.664708 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:35 crc kubenswrapper[4813]: I0317 09:10:35.716133 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:35 crc kubenswrapper[4813]: E0317 09:10:35.716736 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 17 09:10:35 crc kubenswrapper[4813]: I0317 09:10:35.717640 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:35 crc kubenswrapper[4813]: I0317 09:10:35.717719 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:35 crc kubenswrapper[4813]: I0317 09:10:35.717739 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:35 crc kubenswrapper[4813]: I0317 09:10:35.717777 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 09:10:35 crc kubenswrapper[4813]: E0317 09:10:35.724099 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 17 09:10:36 crc kubenswrapper[4813]: I0317 09:10:36.217732 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:10:36 crc kubenswrapper[4813]: I0317 09:10:36.218684 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:36 crc kubenswrapper[4813]: I0317 09:10:36.221248 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:36 crc kubenswrapper[4813]: I0317 09:10:36.221312 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:36 crc kubenswrapper[4813]: I0317 09:10:36.221330 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:36 crc kubenswrapper[4813]: I0317 09:10:36.224451 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:10:36 crc kubenswrapper[4813]: I0317 09:10:36.663551 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:37 crc kubenswrapper[4813]: I0317 09:10:37.008958 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:37 crc kubenswrapper[4813]: I0317 09:10:37.010182 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:37 crc kubenswrapper[4813]: I0317 09:10:37.010253 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:37 crc kubenswrapper[4813]: I0317 09:10:37.010272 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:37 crc kubenswrapper[4813]: I0317 09:10:37.662374 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:38 crc kubenswrapper[4813]: I0317 09:10:38.661788 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:38 crc kubenswrapper[4813]: E0317 09:10:38.822710 4813 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 09:10:39 crc kubenswrapper[4813]: I0317 09:10:39.664389 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:40 crc kubenswrapper[4813]: I0317 09:10:40.661747 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:41 crc kubenswrapper[4813]: I0317 09:10:41.677270 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:42 crc kubenswrapper[4813]: I0317 09:10:42.664309 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:42 crc kubenswrapper[4813]: E0317 09:10:42.722573 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 17 09:10:42 crc kubenswrapper[4813]: I0317 09:10:42.724643 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:42 crc kubenswrapper[4813]: I0317 09:10:42.725886 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:42 crc kubenswrapper[4813]: I0317 09:10:42.725916 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:42 crc kubenswrapper[4813]: I0317 09:10:42.725927 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:42 crc kubenswrapper[4813]: I0317 09:10:42.725947 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 09:10:42 crc kubenswrapper[4813]: E0317 09:10:42.730479 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 17 09:10:43 crc kubenswrapper[4813]: I0317 09:10:43.663757 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:44 crc kubenswrapper[4813]: I0317 09:10:44.660735 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:45 crc kubenswrapper[4813]: I0317 09:10:45.661330 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:46 crc kubenswrapper[4813]: I0317 09:10:46.662304 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:47 crc kubenswrapper[4813]: I0317 09:10:47.667425 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:47 crc kubenswrapper[4813]: I0317 09:10:47.729844 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:47 crc kubenswrapper[4813]: I0317 09:10:47.731424 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:47 crc kubenswrapper[4813]: I0317 09:10:47.731481 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:47 crc kubenswrapper[4813]: I0317 09:10:47.731498 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:47 crc kubenswrapper[4813]: I0317 09:10:47.732285 4813 scope.go:117] "RemoveContainer" containerID="056a72ec3942647bc662c27e16e4d06a6925fe85f46704866499d192b596af48" Mar 17 09:10:48 crc kubenswrapper[4813]: I0317 09:10:48.049069 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 17 09:10:48 crc kubenswrapper[4813]: I0317 09:10:48.051065 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a14d9e3f094ad9e9258a200e9020e7a49b82e421c19f2de3626fa32ba21a44e5"} Mar 17 09:10:48 crc kubenswrapper[4813]: I0317 09:10:48.051203 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:48 crc kubenswrapper[4813]: I0317 09:10:48.052202 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:48 crc kubenswrapper[4813]: I0317 09:10:48.052331 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:48 crc kubenswrapper[4813]: I0317 09:10:48.052418 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:48 crc kubenswrapper[4813]: I0317 09:10:48.661967 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:48 crc kubenswrapper[4813]: E0317 09:10:48.822861 4813 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 09:10:49 crc kubenswrapper[4813]: I0317 09:10:49.565703 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:10:49 crc kubenswrapper[4813]: I0317 09:10:49.566123 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:49 crc kubenswrapper[4813]: I0317 09:10:49.567911 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:49 crc kubenswrapper[4813]: I0317 09:10:49.567945 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:49 crc kubenswrapper[4813]: I0317 09:10:49.567954 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:49 crc kubenswrapper[4813]: I0317 09:10:49.665560 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:49 crc kubenswrapper[4813]: E0317 09:10:49.729145 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 17 09:10:49 crc kubenswrapper[4813]: I0317 09:10:49.731329 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:49 crc kubenswrapper[4813]: I0317 09:10:49.733255 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:49 crc kubenswrapper[4813]: I0317 09:10:49.733306 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:49 crc kubenswrapper[4813]: I0317 09:10:49.733359 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:49 crc kubenswrapper[4813]: I0317 09:10:49.733393 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 09:10:49 crc kubenswrapper[4813]: E0317 09:10:49.737783 4813 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 17 09:10:50 crc kubenswrapper[4813]: I0317 09:10:50.057360 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 17 09:10:50 crc kubenswrapper[4813]: I0317 09:10:50.058044 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 17 09:10:50 crc kubenswrapper[4813]: I0317 09:10:50.060520 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a14d9e3f094ad9e9258a200e9020e7a49b82e421c19f2de3626fa32ba21a44e5" exitCode=255 Mar 17 09:10:50 crc kubenswrapper[4813]: I0317 09:10:50.060560 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a14d9e3f094ad9e9258a200e9020e7a49b82e421c19f2de3626fa32ba21a44e5"} Mar 17 09:10:50 crc kubenswrapper[4813]: I0317 09:10:50.060623 4813 scope.go:117] "RemoveContainer" containerID="056a72ec3942647bc662c27e16e4d06a6925fe85f46704866499d192b596af48" Mar 17 09:10:50 crc kubenswrapper[4813]: I0317 09:10:50.065512 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:50 crc kubenswrapper[4813]: I0317 09:10:50.066406 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:50 crc kubenswrapper[4813]: I0317 09:10:50.066426 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:50 crc kubenswrapper[4813]: I0317 09:10:50.066438 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:50 crc kubenswrapper[4813]: I0317 09:10:50.067238 4813 scope.go:117] "RemoveContainer" containerID="a14d9e3f094ad9e9258a200e9020e7a49b82e421c19f2de3626fa32ba21a44e5" Mar 17 09:10:50 crc kubenswrapper[4813]: E0317 09:10:50.067490 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:10:50 crc kubenswrapper[4813]: W0317 09:10:50.616083 4813 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 17 09:10:50 crc kubenswrapper[4813]: E0317 09:10:50.616133 4813 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 17 09:10:50 crc kubenswrapper[4813]: I0317 09:10:50.665273 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:51 crc kubenswrapper[4813]: I0317 09:10:51.064084 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 17 09:10:51 crc kubenswrapper[4813]: I0317 09:10:51.662727 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:52 crc kubenswrapper[4813]: I0317 09:10:52.083435 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:10:52 crc kubenswrapper[4813]: I0317 09:10:52.083624 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:52 crc kubenswrapper[4813]: I0317 09:10:52.084813 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:52 crc kubenswrapper[4813]: I0317 09:10:52.084848 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:52 crc kubenswrapper[4813]: I0317 09:10:52.084860 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:52 crc kubenswrapper[4813]: I0317 09:10:52.085426 4813 scope.go:117] "RemoveContainer" containerID="a14d9e3f094ad9e9258a200e9020e7a49b82e421c19f2de3626fa32ba21a44e5" Mar 17 09:10:52 crc kubenswrapper[4813]: E0317 09:10:52.085623 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:10:52 crc kubenswrapper[4813]: I0317 09:10:52.371250 4813 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 17 09:10:52 crc kubenswrapper[4813]: I0317 09:10:52.392787 4813 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 17 09:10:52 crc kubenswrapper[4813]: I0317 09:10:52.661982 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:53 crc kubenswrapper[4813]: I0317 09:10:53.667549 4813 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 17 09:10:53 crc kubenswrapper[4813]: I0317 09:10:53.839822 4813 csr.go:261] certificate signing request csr-vlhr2 is approved, waiting to be issued Mar 17 09:10:53 crc kubenswrapper[4813]: I0317 09:10:53.852071 4813 csr.go:257] certificate signing request csr-vlhr2 is issued Mar 17 09:10:53 crc kubenswrapper[4813]: I0317 09:10:53.883319 4813 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 17 09:10:54 crc kubenswrapper[4813]: I0317 09:10:54.446114 4813 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 17 09:10:54 crc kubenswrapper[4813]: I0317 09:10:54.666259 4813 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 17 09:10:54 crc kubenswrapper[4813]: I0317 09:10:54.853952 4813 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2027-01-17 08:03:57.458248132 +0000 UTC Mar 17 09:10:54 crc kubenswrapper[4813]: I0317 09:10:54.854021 4813 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 7342h53m2.604234278s for next certificate rotation Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.738150 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.740317 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.740405 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.740433 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.740590 4813 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.750814 4813 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.751165 4813 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 17 09:10:56 crc kubenswrapper[4813]: E0317 09:10:56.751201 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.755353 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.755423 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.755449 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.755476 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.755494 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:10:56Z","lastTransitionTime":"2026-03-17T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:10:56 crc kubenswrapper[4813]: E0317 09:10:56.774831 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.779929 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.780033 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.780056 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.780084 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.780109 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:10:56Z","lastTransitionTime":"2026-03-17T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:10:56 crc kubenswrapper[4813]: E0317 09:10:56.799520 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.804281 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.804342 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.804360 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.804383 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.804400 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:10:56Z","lastTransitionTime":"2026-03-17T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:10:56 crc kubenswrapper[4813]: E0317 09:10:56.849361 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.857271 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.857306 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.857359 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.857375 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:10:56 crc kubenswrapper[4813]: I0317 09:10:56.857387 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:10:56Z","lastTransitionTime":"2026-03-17T09:10:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:10:56 crc kubenswrapper[4813]: E0317 09:10:56.876822 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:56Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:10:56 crc kubenswrapper[4813]: E0317 09:10:56.876935 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 17 09:10:56 crc kubenswrapper[4813]: E0317 09:10:56.876963 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:56 crc kubenswrapper[4813]: E0317 09:10:56.978068 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:57 crc kubenswrapper[4813]: E0317 09:10:57.078804 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:57 crc kubenswrapper[4813]: E0317 09:10:57.179960 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:57 crc kubenswrapper[4813]: E0317 09:10:57.280392 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:57 crc kubenswrapper[4813]: E0317 09:10:57.381078 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:57 crc kubenswrapper[4813]: E0317 09:10:57.482277 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:57 crc kubenswrapper[4813]: E0317 09:10:57.582642 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:57 crc kubenswrapper[4813]: E0317 09:10:57.683554 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:57 crc kubenswrapper[4813]: E0317 09:10:57.784504 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:57 crc kubenswrapper[4813]: E0317 09:10:57.885111 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:57 crc kubenswrapper[4813]: E0317 09:10:57.985688 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:58 crc kubenswrapper[4813]: E0317 09:10:58.086753 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:58 crc kubenswrapper[4813]: E0317 09:10:58.187840 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:58 crc kubenswrapper[4813]: E0317 09:10:58.288225 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:58 crc kubenswrapper[4813]: E0317 09:10:58.388573 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:58 crc kubenswrapper[4813]: E0317 09:10:58.489684 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:58 crc kubenswrapper[4813]: E0317 09:10:58.590165 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:58 crc kubenswrapper[4813]: E0317 09:10:58.691254 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:58 crc kubenswrapper[4813]: E0317 09:10:58.792104 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:58 crc kubenswrapper[4813]: E0317 09:10:58.823730 4813 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 09:10:58 crc kubenswrapper[4813]: E0317 09:10:58.892877 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:58 crc kubenswrapper[4813]: E0317 09:10:58.993187 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:59 crc kubenswrapper[4813]: E0317 09:10:59.093909 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:59 crc kubenswrapper[4813]: E0317 09:10:59.194989 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:59 crc kubenswrapper[4813]: E0317 09:10:59.295645 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:59 crc kubenswrapper[4813]: E0317 09:10:59.396829 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:59 crc kubenswrapper[4813]: E0317 09:10:59.497924 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:59 crc kubenswrapper[4813]: E0317 09:10:59.598489 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:59 crc kubenswrapper[4813]: E0317 09:10:59.699686 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:59 crc kubenswrapper[4813]: E0317 09:10:59.799791 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:10:59 crc kubenswrapper[4813]: E0317 09:10:59.901173 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:00 crc kubenswrapper[4813]: E0317 09:11:00.001818 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:00 crc kubenswrapper[4813]: E0317 09:11:00.101994 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:00 crc kubenswrapper[4813]: E0317 09:11:00.202560 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:00 crc kubenswrapper[4813]: E0317 09:11:00.302705 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:00 crc kubenswrapper[4813]: E0317 09:11:00.403272 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:00 crc kubenswrapper[4813]: E0317 09:11:00.504130 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:00 crc kubenswrapper[4813]: E0317 09:11:00.605241 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:00 crc kubenswrapper[4813]: E0317 09:11:00.706401 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:00 crc kubenswrapper[4813]: E0317 09:11:00.806695 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:00 crc kubenswrapper[4813]: E0317 09:11:00.906899 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:01 crc kubenswrapper[4813]: E0317 09:11:01.007381 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:01 crc kubenswrapper[4813]: E0317 09:11:01.108308 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:01 crc kubenswrapper[4813]: E0317 09:11:01.208723 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:01 crc kubenswrapper[4813]: E0317 09:11:01.309261 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:01 crc kubenswrapper[4813]: E0317 09:11:01.409828 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:01 crc kubenswrapper[4813]: E0317 09:11:01.510377 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:01 crc kubenswrapper[4813]: E0317 09:11:01.611220 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:01 crc kubenswrapper[4813]: E0317 09:11:01.712164 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:01 crc kubenswrapper[4813]: E0317 09:11:01.812795 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:01 crc kubenswrapper[4813]: E0317 09:11:01.913297 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:02 crc kubenswrapper[4813]: E0317 09:11:02.014018 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:02 crc kubenswrapper[4813]: E0317 09:11:02.114741 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:02 crc kubenswrapper[4813]: E0317 09:11:02.216202 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:02 crc kubenswrapper[4813]: E0317 09:11:02.316757 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:02 crc kubenswrapper[4813]: E0317 09:11:02.416900 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:02 crc kubenswrapper[4813]: E0317 09:11:02.517436 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:02 crc kubenswrapper[4813]: E0317 09:11:02.618591 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:02 crc kubenswrapper[4813]: E0317 09:11:02.719140 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:02 crc kubenswrapper[4813]: E0317 09:11:02.820080 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:02 crc kubenswrapper[4813]: E0317 09:11:02.920680 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:03 crc kubenswrapper[4813]: E0317 09:11:03.021397 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:03 crc kubenswrapper[4813]: E0317 09:11:03.122444 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:03 crc kubenswrapper[4813]: E0317 09:11:03.222764 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:03 crc kubenswrapper[4813]: E0317 09:11:03.323561 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:03 crc kubenswrapper[4813]: E0317 09:11:03.424570 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:03 crc kubenswrapper[4813]: E0317 09:11:03.525685 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:03 crc kubenswrapper[4813]: E0317 09:11:03.626099 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:03 crc kubenswrapper[4813]: E0317 09:11:03.726285 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:03 crc kubenswrapper[4813]: I0317 09:11:03.730101 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:11:03 crc kubenswrapper[4813]: I0317 09:11:03.732400 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:03 crc kubenswrapper[4813]: I0317 09:11:03.732474 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:03 crc kubenswrapper[4813]: I0317 09:11:03.732499 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:03 crc kubenswrapper[4813]: E0317 09:11:03.827036 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:03 crc kubenswrapper[4813]: E0317 09:11:03.928033 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:04 crc kubenswrapper[4813]: E0317 09:11:04.028433 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:04 crc kubenswrapper[4813]: E0317 09:11:04.129216 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:04 crc kubenswrapper[4813]: E0317 09:11:04.229673 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:04 crc kubenswrapper[4813]: E0317 09:11:04.330277 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:04 crc kubenswrapper[4813]: E0317 09:11:04.430902 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:04 crc kubenswrapper[4813]: E0317 09:11:04.531052 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:04 crc kubenswrapper[4813]: E0317 09:11:04.632220 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:04 crc kubenswrapper[4813]: I0317 09:11:04.730543 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:11:04 crc kubenswrapper[4813]: I0317 09:11:04.732178 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:04 crc kubenswrapper[4813]: I0317 09:11:04.732235 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:04 crc kubenswrapper[4813]: I0317 09:11:04.732255 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:04 crc kubenswrapper[4813]: E0317 09:11:04.732313 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:04 crc kubenswrapper[4813]: I0317 09:11:04.733197 4813 scope.go:117] "RemoveContainer" containerID="a14d9e3f094ad9e9258a200e9020e7a49b82e421c19f2de3626fa32ba21a44e5" Mar 17 09:11:04 crc kubenswrapper[4813]: E0317 09:11:04.733524 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:11:04 crc kubenswrapper[4813]: E0317 09:11:04.832820 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:04 crc kubenswrapper[4813]: E0317 09:11:04.934230 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:05 crc kubenswrapper[4813]: E0317 09:11:05.034379 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:05 crc kubenswrapper[4813]: E0317 09:11:05.134961 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:05 crc kubenswrapper[4813]: E0317 09:11:05.236303 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:05 crc kubenswrapper[4813]: E0317 09:11:05.336929 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:05 crc kubenswrapper[4813]: E0317 09:11:05.438140 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:05 crc kubenswrapper[4813]: E0317 09:11:05.539053 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:05 crc kubenswrapper[4813]: E0317 09:11:05.639981 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:05 crc kubenswrapper[4813]: E0317 09:11:05.740486 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:05 crc kubenswrapper[4813]: E0317 09:11:05.841718 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:05 crc kubenswrapper[4813]: E0317 09:11:05.942492 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:06 crc kubenswrapper[4813]: E0317 09:11:06.043359 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:06 crc kubenswrapper[4813]: E0317 09:11:06.144510 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:06 crc kubenswrapper[4813]: E0317 09:11:06.245591 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:06 crc kubenswrapper[4813]: E0317 09:11:06.345994 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:06 crc kubenswrapper[4813]: E0317 09:11:06.446510 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:06 crc kubenswrapper[4813]: E0317 09:11:06.547722 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:06 crc kubenswrapper[4813]: E0317 09:11:06.649075 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:06 crc kubenswrapper[4813]: E0317 09:11:06.749391 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:06 crc kubenswrapper[4813]: E0317 09:11:06.849800 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:06 crc kubenswrapper[4813]: E0317 09:11:06.950575 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:07 crc kubenswrapper[4813]: E0317 09:11:07.050854 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:07 crc kubenswrapper[4813]: E0317 09:11:07.151305 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:07 crc kubenswrapper[4813]: E0317 09:11:07.240916 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 17 09:11:07 crc kubenswrapper[4813]: I0317 09:11:07.245224 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:07 crc kubenswrapper[4813]: I0317 09:11:07.245491 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:07 crc kubenswrapper[4813]: I0317 09:11:07.245680 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:07 crc kubenswrapper[4813]: I0317 09:11:07.245843 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:07 crc kubenswrapper[4813]: I0317 09:11:07.245982 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:07Z","lastTransitionTime":"2026-03-17T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:07 crc kubenswrapper[4813]: E0317 09:11:07.258327 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:07 crc kubenswrapper[4813]: I0317 09:11:07.263667 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:07 crc kubenswrapper[4813]: I0317 09:11:07.263978 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:07 crc kubenswrapper[4813]: I0317 09:11:07.264215 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:07 crc kubenswrapper[4813]: I0317 09:11:07.264429 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:07 crc kubenswrapper[4813]: I0317 09:11:07.264636 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:07Z","lastTransitionTime":"2026-03-17T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:07 crc kubenswrapper[4813]: E0317 09:11:07.280958 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:07 crc kubenswrapper[4813]: I0317 09:11:07.285465 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:07 crc kubenswrapper[4813]: I0317 09:11:07.285518 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:07 crc kubenswrapper[4813]: I0317 09:11:07.285536 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:07 crc kubenswrapper[4813]: I0317 09:11:07.285559 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:07 crc kubenswrapper[4813]: I0317 09:11:07.285577 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:07Z","lastTransitionTime":"2026-03-17T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:07 crc kubenswrapper[4813]: E0317 09:11:07.297777 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:07 crc kubenswrapper[4813]: I0317 09:11:07.301562 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:07 crc kubenswrapper[4813]: I0317 09:11:07.301608 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:07 crc kubenswrapper[4813]: I0317 09:11:07.301627 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:07 crc kubenswrapper[4813]: I0317 09:11:07.301645 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:07 crc kubenswrapper[4813]: I0317 09:11:07.301655 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:07Z","lastTransitionTime":"2026-03-17T09:11:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:07 crc kubenswrapper[4813]: E0317 09:11:07.315490 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:07 crc kubenswrapper[4813]: E0317 09:11:07.315788 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 17 09:11:07 crc kubenswrapper[4813]: E0317 09:11:07.315830 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:07 crc kubenswrapper[4813]: E0317 09:11:07.416258 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:07 crc kubenswrapper[4813]: E0317 09:11:07.516937 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:07 crc kubenswrapper[4813]: E0317 09:11:07.617521 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:07 crc kubenswrapper[4813]: E0317 09:11:07.718432 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:07 crc kubenswrapper[4813]: E0317 09:11:07.818957 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:07 crc kubenswrapper[4813]: E0317 09:11:07.919481 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:08 crc kubenswrapper[4813]: E0317 09:11:08.020275 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:08 crc kubenswrapper[4813]: E0317 09:11:08.120950 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:08 crc kubenswrapper[4813]: E0317 09:11:08.222049 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:08 crc kubenswrapper[4813]: E0317 09:11:08.322856 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:08 crc kubenswrapper[4813]: E0317 09:11:08.423469 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:08 crc kubenswrapper[4813]: E0317 09:11:08.524707 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:08 crc kubenswrapper[4813]: E0317 09:11:08.625762 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:08 crc kubenswrapper[4813]: E0317 09:11:08.726227 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:08 crc kubenswrapper[4813]: E0317 09:11:08.824783 4813 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 09:11:08 crc kubenswrapper[4813]: E0317 09:11:08.827046 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:08 crc kubenswrapper[4813]: E0317 09:11:08.927443 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:09 crc kubenswrapper[4813]: E0317 09:11:09.028038 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:09 crc kubenswrapper[4813]: E0317 09:11:09.128925 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:09 crc kubenswrapper[4813]: E0317 09:11:09.229771 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:09 crc kubenswrapper[4813]: E0317 09:11:09.330190 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:09 crc kubenswrapper[4813]: E0317 09:11:09.430453 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:09 crc kubenswrapper[4813]: E0317 09:11:09.531170 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:09 crc kubenswrapper[4813]: E0317 09:11:09.631294 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:09 crc kubenswrapper[4813]: E0317 09:11:09.731390 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:09 crc kubenswrapper[4813]: E0317 09:11:09.831986 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:09 crc kubenswrapper[4813]: E0317 09:11:09.932467 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:10 crc kubenswrapper[4813]: E0317 09:11:10.042416 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:10 crc kubenswrapper[4813]: E0317 09:11:10.143146 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:10 crc kubenswrapper[4813]: E0317 09:11:10.244263 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:10 crc kubenswrapper[4813]: E0317 09:11:10.345290 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:10 crc kubenswrapper[4813]: E0317 09:11:10.446375 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:10 crc kubenswrapper[4813]: E0317 09:11:10.546611 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:10 crc kubenswrapper[4813]: E0317 09:11:10.647186 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:10 crc kubenswrapper[4813]: E0317 09:11:10.748096 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:10 crc kubenswrapper[4813]: E0317 09:11:10.849026 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:10 crc kubenswrapper[4813]: E0317 09:11:10.950127 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:11 crc kubenswrapper[4813]: E0317 09:11:11.050513 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:11 crc kubenswrapper[4813]: E0317 09:11:11.151310 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:11 crc kubenswrapper[4813]: E0317 09:11:11.252087 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:11 crc kubenswrapper[4813]: E0317 09:11:11.353174 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:11 crc kubenswrapper[4813]: E0317 09:11:11.453312 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:11 crc kubenswrapper[4813]: I0317 09:11:11.521389 4813 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 17 09:11:11 crc kubenswrapper[4813]: E0317 09:11:11.554456 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:11 crc kubenswrapper[4813]: E0317 09:11:11.654635 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:11 crc kubenswrapper[4813]: E0317 09:11:11.755035 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:11 crc kubenswrapper[4813]: E0317 09:11:11.856069 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:11 crc kubenswrapper[4813]: E0317 09:11:11.956836 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:12 crc kubenswrapper[4813]: E0317 09:11:12.057853 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:12 crc kubenswrapper[4813]: E0317 09:11:12.158485 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:12 crc kubenswrapper[4813]: E0317 09:11:12.259030 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:12 crc kubenswrapper[4813]: E0317 09:11:12.360078 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:12 crc kubenswrapper[4813]: E0317 09:11:12.460464 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:12 crc kubenswrapper[4813]: E0317 09:11:12.560833 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:12 crc kubenswrapper[4813]: E0317 09:11:12.661238 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:12 crc kubenswrapper[4813]: E0317 09:11:12.762283 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:12 crc kubenswrapper[4813]: E0317 09:11:12.863242 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:12 crc kubenswrapper[4813]: E0317 09:11:12.963462 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:13 crc kubenswrapper[4813]: E0317 09:11:13.064036 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:13 crc kubenswrapper[4813]: E0317 09:11:13.165053 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:13 crc kubenswrapper[4813]: E0317 09:11:13.265187 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:13 crc kubenswrapper[4813]: E0317 09:11:13.365325 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:13 crc kubenswrapper[4813]: E0317 09:11:13.465644 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:13 crc kubenswrapper[4813]: E0317 09:11:13.566580 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:13 crc kubenswrapper[4813]: E0317 09:11:13.667812 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:13 crc kubenswrapper[4813]: E0317 09:11:13.768458 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:13 crc kubenswrapper[4813]: E0317 09:11:13.868743 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:13 crc kubenswrapper[4813]: E0317 09:11:13.969703 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:14 crc kubenswrapper[4813]: E0317 09:11:14.070296 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:14 crc kubenswrapper[4813]: E0317 09:11:14.171261 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:14 crc kubenswrapper[4813]: E0317 09:11:14.272419 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:14 crc kubenswrapper[4813]: E0317 09:11:14.372863 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:14 crc kubenswrapper[4813]: E0317 09:11:14.473003 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:14 crc kubenswrapper[4813]: I0317 09:11:14.498018 4813 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 17 09:11:14 crc kubenswrapper[4813]: E0317 09:11:14.573775 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:14 crc kubenswrapper[4813]: E0317 09:11:14.674438 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:14 crc kubenswrapper[4813]: E0317 09:11:14.775158 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:14 crc kubenswrapper[4813]: E0317 09:11:14.875368 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:14 crc kubenswrapper[4813]: E0317 09:11:14.976835 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:15 crc kubenswrapper[4813]: E0317 09:11:15.077782 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:15 crc kubenswrapper[4813]: E0317 09:11:15.178400 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:15 crc kubenswrapper[4813]: E0317 09:11:15.279509 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:15 crc kubenswrapper[4813]: E0317 09:11:15.380576 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:15 crc kubenswrapper[4813]: E0317 09:11:15.481288 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:15 crc kubenswrapper[4813]: E0317 09:11:15.581689 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:15 crc kubenswrapper[4813]: E0317 09:11:15.682904 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:15 crc kubenswrapper[4813]: E0317 09:11:15.783363 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:15 crc kubenswrapper[4813]: E0317 09:11:15.883714 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:15 crc kubenswrapper[4813]: E0317 09:11:15.983914 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:16 crc kubenswrapper[4813]: E0317 09:11:16.085365 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:16 crc kubenswrapper[4813]: E0317 09:11:16.185839 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:16 crc kubenswrapper[4813]: E0317 09:11:16.286238 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:16 crc kubenswrapper[4813]: E0317 09:11:16.386515 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:16 crc kubenswrapper[4813]: E0317 09:11:16.487675 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:16 crc kubenswrapper[4813]: E0317 09:11:16.587945 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:16 crc kubenswrapper[4813]: E0317 09:11:16.688778 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:16 crc kubenswrapper[4813]: I0317 09:11:16.729683 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:11:16 crc kubenswrapper[4813]: I0317 09:11:16.730793 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:16 crc kubenswrapper[4813]: I0317 09:11:16.730997 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:16 crc kubenswrapper[4813]: I0317 09:11:16.731145 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:16 crc kubenswrapper[4813]: I0317 09:11:16.732270 4813 scope.go:117] "RemoveContainer" containerID="a14d9e3f094ad9e9258a200e9020e7a49b82e421c19f2de3626fa32ba21a44e5" Mar 17 09:11:16 crc kubenswrapper[4813]: E0317 09:11:16.732774 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:11:16 crc kubenswrapper[4813]: E0317 09:11:16.789649 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:16 crc kubenswrapper[4813]: E0317 09:11:16.890211 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:16 crc kubenswrapper[4813]: E0317 09:11:16.990738 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:17 crc kubenswrapper[4813]: E0317 09:11:17.091353 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:17 crc kubenswrapper[4813]: E0317 09:11:17.192106 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:17 crc kubenswrapper[4813]: E0317 09:11:17.293509 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:17 crc kubenswrapper[4813]: E0317 09:11:17.394433 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:17 crc kubenswrapper[4813]: E0317 09:11:17.495007 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:17 crc kubenswrapper[4813]: E0317 09:11:17.596087 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:17 crc kubenswrapper[4813]: E0317 09:11:17.679011 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 17 09:11:17 crc kubenswrapper[4813]: I0317 09:11:17.684473 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:17 crc kubenswrapper[4813]: I0317 09:11:17.684548 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:17 crc kubenswrapper[4813]: I0317 09:11:17.684573 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:17 crc kubenswrapper[4813]: I0317 09:11:17.684634 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:17 crc kubenswrapper[4813]: I0317 09:11:17.684663 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:17Z","lastTransitionTime":"2026-03-17T09:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:17 crc kubenswrapper[4813]: E0317 09:11:17.700728 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:17 crc kubenswrapper[4813]: I0317 09:11:17.705555 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:17 crc kubenswrapper[4813]: I0317 09:11:17.705639 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:17 crc kubenswrapper[4813]: I0317 09:11:17.705659 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:17 crc kubenswrapper[4813]: I0317 09:11:17.705683 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:17 crc kubenswrapper[4813]: I0317 09:11:17.705704 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:17Z","lastTransitionTime":"2026-03-17T09:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:17 crc kubenswrapper[4813]: E0317 09:11:17.721768 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:17 crc kubenswrapper[4813]: I0317 09:11:17.726166 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:17 crc kubenswrapper[4813]: I0317 09:11:17.726222 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:17 crc kubenswrapper[4813]: I0317 09:11:17.726240 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:17 crc kubenswrapper[4813]: I0317 09:11:17.726261 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:17 crc kubenswrapper[4813]: I0317 09:11:17.726278 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:17Z","lastTransitionTime":"2026-03-17T09:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:17 crc kubenswrapper[4813]: E0317 09:11:17.742313 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:17 crc kubenswrapper[4813]: I0317 09:11:17.746748 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:17 crc kubenswrapper[4813]: I0317 09:11:17.747007 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:17 crc kubenswrapper[4813]: I0317 09:11:17.747168 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:17 crc kubenswrapper[4813]: I0317 09:11:17.747327 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:17 crc kubenswrapper[4813]: I0317 09:11:17.747465 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:17Z","lastTransitionTime":"2026-03-17T09:11:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:17 crc kubenswrapper[4813]: E0317 09:11:17.771585 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:17 crc kubenswrapper[4813]: E0317 09:11:17.771969 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 17 09:11:17 crc kubenswrapper[4813]: E0317 09:11:17.772021 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:17 crc kubenswrapper[4813]: E0317 09:11:17.872637 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:17 crc kubenswrapper[4813]: E0317 09:11:17.973424 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:18 crc kubenswrapper[4813]: E0317 09:11:18.074058 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:18 crc kubenswrapper[4813]: E0317 09:11:18.174944 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:18 crc kubenswrapper[4813]: E0317 09:11:18.275071 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:18 crc kubenswrapper[4813]: E0317 09:11:18.375517 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:18 crc kubenswrapper[4813]: E0317 09:11:18.476268 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:18 crc kubenswrapper[4813]: E0317 09:11:18.576960 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:18 crc kubenswrapper[4813]: E0317 09:11:18.677391 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:18 crc kubenswrapper[4813]: E0317 09:11:18.778570 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:18 crc kubenswrapper[4813]: E0317 09:11:18.825914 4813 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 09:11:18 crc kubenswrapper[4813]: E0317 09:11:18.879043 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:18 crc kubenswrapper[4813]: E0317 09:11:18.980126 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:19 crc kubenswrapper[4813]: E0317 09:11:19.080317 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:19 crc kubenswrapper[4813]: E0317 09:11:19.180882 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:19 crc kubenswrapper[4813]: E0317 09:11:19.281245 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:19 crc kubenswrapper[4813]: E0317 09:11:19.382312 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:19 crc kubenswrapper[4813]: E0317 09:11:19.482704 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:19 crc kubenswrapper[4813]: E0317 09:11:19.583314 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:19 crc kubenswrapper[4813]: E0317 09:11:19.684506 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:19 crc kubenswrapper[4813]: E0317 09:11:19.785456 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:19 crc kubenswrapper[4813]: E0317 09:11:19.886200 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:19 crc kubenswrapper[4813]: E0317 09:11:19.986317 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:20 crc kubenswrapper[4813]: E0317 09:11:20.086960 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:20 crc kubenswrapper[4813]: E0317 09:11:20.188051 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:20 crc kubenswrapper[4813]: E0317 09:11:20.289125 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:20 crc kubenswrapper[4813]: E0317 09:11:20.389474 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:20 crc kubenswrapper[4813]: E0317 09:11:20.490644 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:20 crc kubenswrapper[4813]: E0317 09:11:20.590761 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:20 crc kubenswrapper[4813]: E0317 09:11:20.691268 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:20 crc kubenswrapper[4813]: E0317 09:11:20.791969 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:20 crc kubenswrapper[4813]: E0317 09:11:20.892943 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:20 crc kubenswrapper[4813]: E0317 09:11:20.993082 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:21 crc kubenswrapper[4813]: E0317 09:11:21.093264 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:21 crc kubenswrapper[4813]: E0317 09:11:21.194119 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:21 crc kubenswrapper[4813]: E0317 09:11:21.295258 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:21 crc kubenswrapper[4813]: E0317 09:11:21.396128 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:21 crc kubenswrapper[4813]: E0317 09:11:21.496252 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:21 crc kubenswrapper[4813]: E0317 09:11:21.597337 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:21 crc kubenswrapper[4813]: E0317 09:11:21.697854 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:21 crc kubenswrapper[4813]: E0317 09:11:21.798076 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:21 crc kubenswrapper[4813]: E0317 09:11:21.898910 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:21 crc kubenswrapper[4813]: E0317 09:11:21.999262 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:22 crc kubenswrapper[4813]: E0317 09:11:22.100349 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:22 crc kubenswrapper[4813]: E0317 09:11:22.200785 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:22 crc kubenswrapper[4813]: E0317 09:11:22.300982 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:22 crc kubenswrapper[4813]: E0317 09:11:22.402096 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:22 crc kubenswrapper[4813]: E0317 09:11:22.502195 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:22 crc kubenswrapper[4813]: E0317 09:11:22.603007 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:22 crc kubenswrapper[4813]: E0317 09:11:22.703977 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:22 crc kubenswrapper[4813]: E0317 09:11:22.804954 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:22 crc kubenswrapper[4813]: E0317 09:11:22.905588 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:23 crc kubenswrapper[4813]: E0317 09:11:23.005951 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:23 crc kubenswrapper[4813]: E0317 09:11:23.107007 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:23 crc kubenswrapper[4813]: E0317 09:11:23.207872 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:23 crc kubenswrapper[4813]: E0317 09:11:23.308446 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:23 crc kubenswrapper[4813]: E0317 09:11:23.408759 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:23 crc kubenswrapper[4813]: E0317 09:11:23.509261 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:23 crc kubenswrapper[4813]: E0317 09:11:23.609922 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:23 crc kubenswrapper[4813]: E0317 09:11:23.710121 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:23 crc kubenswrapper[4813]: I0317 09:11:23.729897 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:11:23 crc kubenswrapper[4813]: I0317 09:11:23.731437 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:23 crc kubenswrapper[4813]: I0317 09:11:23.731494 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:23 crc kubenswrapper[4813]: I0317 09:11:23.731512 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:23 crc kubenswrapper[4813]: E0317 09:11:23.811253 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:23 crc kubenswrapper[4813]: E0317 09:11:23.911942 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:24 crc kubenswrapper[4813]: E0317 09:11:24.012378 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:24 crc kubenswrapper[4813]: E0317 09:11:24.113419 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:24 crc kubenswrapper[4813]: E0317 09:11:24.214440 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:24 crc kubenswrapper[4813]: E0317 09:11:24.315158 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:24 crc kubenswrapper[4813]: E0317 09:11:24.415594 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:24 crc kubenswrapper[4813]: E0317 09:11:24.516327 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:24 crc kubenswrapper[4813]: E0317 09:11:24.617495 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:24 crc kubenswrapper[4813]: E0317 09:11:24.718001 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:24 crc kubenswrapper[4813]: E0317 09:11:24.818998 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:24 crc kubenswrapper[4813]: E0317 09:11:24.919636 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:25 crc kubenswrapper[4813]: E0317 09:11:25.020204 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:25 crc kubenswrapper[4813]: E0317 09:11:25.121235 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:25 crc kubenswrapper[4813]: E0317 09:11:25.222122 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:25 crc kubenswrapper[4813]: E0317 09:11:25.323057 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:25 crc kubenswrapper[4813]: E0317 09:11:25.423955 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:25 crc kubenswrapper[4813]: E0317 09:11:25.524210 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:25 crc kubenswrapper[4813]: E0317 09:11:25.624679 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:25 crc kubenswrapper[4813]: E0317 09:11:25.725409 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:25 crc kubenswrapper[4813]: E0317 09:11:25.826032 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:25 crc kubenswrapper[4813]: E0317 09:11:25.926275 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:26 crc kubenswrapper[4813]: E0317 09:11:26.026675 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:26 crc kubenswrapper[4813]: E0317 09:11:26.127203 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:26 crc kubenswrapper[4813]: E0317 09:11:26.227738 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:26 crc kubenswrapper[4813]: E0317 09:11:26.328840 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:26 crc kubenswrapper[4813]: E0317 09:11:26.429954 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:26 crc kubenswrapper[4813]: E0317 09:11:26.530409 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:26 crc kubenswrapper[4813]: E0317 09:11:26.630564 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:26 crc kubenswrapper[4813]: E0317 09:11:26.731668 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:26 crc kubenswrapper[4813]: E0317 09:11:26.832669 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:26 crc kubenswrapper[4813]: E0317 09:11:26.932765 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:27 crc kubenswrapper[4813]: E0317 09:11:27.033713 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:27 crc kubenswrapper[4813]: E0317 09:11:27.134198 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:27 crc kubenswrapper[4813]: E0317 09:11:27.235406 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:27 crc kubenswrapper[4813]: E0317 09:11:27.336553 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:27 crc kubenswrapper[4813]: E0317 09:11:27.437734 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:27 crc kubenswrapper[4813]: E0317 09:11:27.538173 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:27 crc kubenswrapper[4813]: E0317 09:11:27.671984 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.729738 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.730884 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.730950 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.730969 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.732052 4813 scope.go:117] "RemoveContainer" containerID="a14d9e3f094ad9e9258a200e9020e7a49b82e421c19f2de3626fa32ba21a44e5" Mar 17 09:11:27 crc kubenswrapper[4813]: E0317 09:11:27.732411 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:11:27 crc kubenswrapper[4813]: E0317 09:11:27.772947 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:27 crc kubenswrapper[4813]: E0317 09:11:27.792326 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.798161 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.798222 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.798259 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.798291 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.798313 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:27Z","lastTransitionTime":"2026-03-17T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:27 crc kubenswrapper[4813]: E0317 09:11:27.818495 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.824611 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.824730 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.824748 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.824775 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.824793 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:27Z","lastTransitionTime":"2026-03-17T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:27 crc kubenswrapper[4813]: E0317 09:11:27.840740 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.846217 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.846273 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.846300 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.846330 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.846350 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:27Z","lastTransitionTime":"2026-03-17T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:27 crc kubenswrapper[4813]: E0317 09:11:27.861752 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.866721 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.866775 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.866793 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.866817 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:27 crc kubenswrapper[4813]: I0317 09:11:27.866835 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:27Z","lastTransitionTime":"2026-03-17T09:11:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:27 crc kubenswrapper[4813]: E0317 09:11:27.883540 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:27 crc kubenswrapper[4813]: E0317 09:11:27.883958 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 17 09:11:27 crc kubenswrapper[4813]: E0317 09:11:27.884012 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:27 crc kubenswrapper[4813]: E0317 09:11:27.984691 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:28 crc kubenswrapper[4813]: E0317 09:11:28.085674 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:28 crc kubenswrapper[4813]: E0317 09:11:28.186456 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:28 crc kubenswrapper[4813]: E0317 09:11:28.286801 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:28 crc kubenswrapper[4813]: E0317 09:11:28.388094 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:28 crc kubenswrapper[4813]: E0317 09:11:28.489110 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:28 crc kubenswrapper[4813]: E0317 09:11:28.589281 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:28 crc kubenswrapper[4813]: E0317 09:11:28.689593 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:28 crc kubenswrapper[4813]: E0317 09:11:28.790692 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:28 crc kubenswrapper[4813]: E0317 09:11:28.826334 4813 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 09:11:28 crc kubenswrapper[4813]: E0317 09:11:28.891180 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:28 crc kubenswrapper[4813]: E0317 09:11:28.991630 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:29 crc kubenswrapper[4813]: E0317 09:11:29.092439 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:29 crc kubenswrapper[4813]: E0317 09:11:29.193668 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:29 crc kubenswrapper[4813]: E0317 09:11:29.294485 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:29 crc kubenswrapper[4813]: E0317 09:11:29.394977 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:29 crc kubenswrapper[4813]: E0317 09:11:29.496334 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:29 crc kubenswrapper[4813]: E0317 09:11:29.596880 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:29 crc kubenswrapper[4813]: E0317 09:11:29.697287 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:29 crc kubenswrapper[4813]: E0317 09:11:29.797938 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:29 crc kubenswrapper[4813]: E0317 09:11:29.899010 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:29 crc kubenswrapper[4813]: E0317 09:11:29.999564 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:30 crc kubenswrapper[4813]: E0317 09:11:30.100918 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:30 crc kubenswrapper[4813]: E0317 09:11:30.201736 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:30 crc kubenswrapper[4813]: E0317 09:11:30.302451 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:30 crc kubenswrapper[4813]: E0317 09:11:30.403288 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:30 crc kubenswrapper[4813]: E0317 09:11:30.504684 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:30 crc kubenswrapper[4813]: E0317 09:11:30.605819 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:30 crc kubenswrapper[4813]: E0317 09:11:30.706671 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:30 crc kubenswrapper[4813]: E0317 09:11:30.807004 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:30 crc kubenswrapper[4813]: E0317 09:11:30.907529 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:31 crc kubenswrapper[4813]: E0317 09:11:31.008727 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:31 crc kubenswrapper[4813]: E0317 09:11:31.110713 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:31 crc kubenswrapper[4813]: E0317 09:11:31.211396 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:31 crc kubenswrapper[4813]: E0317 09:11:31.313503 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:31 crc kubenswrapper[4813]: E0317 09:11:31.414364 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:31 crc kubenswrapper[4813]: E0317 09:11:31.514875 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:31 crc kubenswrapper[4813]: E0317 09:11:31.615836 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:31 crc kubenswrapper[4813]: E0317 09:11:31.716867 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:31 crc kubenswrapper[4813]: E0317 09:11:31.818083 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:31 crc kubenswrapper[4813]: E0317 09:11:31.918354 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:32 crc kubenswrapper[4813]: E0317 09:11:32.019323 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:32 crc kubenswrapper[4813]: E0317 09:11:32.119891 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:32 crc kubenswrapper[4813]: E0317 09:11:32.220476 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:32 crc kubenswrapper[4813]: E0317 09:11:32.320626 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:32 crc kubenswrapper[4813]: E0317 09:11:32.421070 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:32 crc kubenswrapper[4813]: E0317 09:11:32.521222 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:32 crc kubenswrapper[4813]: E0317 09:11:32.621993 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:32 crc kubenswrapper[4813]: E0317 09:11:32.722946 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:32 crc kubenswrapper[4813]: E0317 09:11:32.823800 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:32 crc kubenswrapper[4813]: E0317 09:11:32.924090 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:33 crc kubenswrapper[4813]: E0317 09:11:33.024936 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:33 crc kubenswrapper[4813]: E0317 09:11:33.125808 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:33 crc kubenswrapper[4813]: E0317 09:11:33.227014 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:33 crc kubenswrapper[4813]: E0317 09:11:33.327829 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:33 crc kubenswrapper[4813]: E0317 09:11:33.428531 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:33 crc kubenswrapper[4813]: E0317 09:11:33.529656 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:33 crc kubenswrapper[4813]: E0317 09:11:33.630062 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:33 crc kubenswrapper[4813]: E0317 09:11:33.730924 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:33 crc kubenswrapper[4813]: E0317 09:11:33.831575 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:33 crc kubenswrapper[4813]: E0317 09:11:33.932283 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:34 crc kubenswrapper[4813]: E0317 09:11:34.033114 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:34 crc kubenswrapper[4813]: E0317 09:11:34.133682 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:34 crc kubenswrapper[4813]: E0317 09:11:34.234770 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:34 crc kubenswrapper[4813]: E0317 09:11:34.335107 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:34 crc kubenswrapper[4813]: E0317 09:11:34.435396 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:34 crc kubenswrapper[4813]: E0317 09:11:34.535850 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:34 crc kubenswrapper[4813]: E0317 09:11:34.636188 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:34 crc kubenswrapper[4813]: E0317 09:11:34.737085 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:34 crc kubenswrapper[4813]: E0317 09:11:34.837972 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:34 crc kubenswrapper[4813]: E0317 09:11:34.939117 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:35 crc kubenswrapper[4813]: E0317 09:11:35.040200 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:35 crc kubenswrapper[4813]: E0317 09:11:35.141111 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:35 crc kubenswrapper[4813]: E0317 09:11:35.241780 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:35 crc kubenswrapper[4813]: E0317 09:11:35.342849 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:35 crc kubenswrapper[4813]: E0317 09:11:35.444070 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:35 crc kubenswrapper[4813]: E0317 09:11:35.544491 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:35 crc kubenswrapper[4813]: E0317 09:11:35.644693 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:35 crc kubenswrapper[4813]: E0317 09:11:35.745399 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:35 crc kubenswrapper[4813]: E0317 09:11:35.846303 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:35 crc kubenswrapper[4813]: E0317 09:11:35.947294 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:36 crc kubenswrapper[4813]: E0317 09:11:36.048256 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:36 crc kubenswrapper[4813]: E0317 09:11:36.149229 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:36 crc kubenswrapper[4813]: E0317 09:11:36.249842 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:36 crc kubenswrapper[4813]: E0317 09:11:36.350879 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:36 crc kubenswrapper[4813]: E0317 09:11:36.451827 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:36 crc kubenswrapper[4813]: E0317 09:11:36.552943 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:36 crc kubenswrapper[4813]: E0317 09:11:36.653414 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:36 crc kubenswrapper[4813]: E0317 09:11:36.753930 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:36 crc kubenswrapper[4813]: E0317 09:11:36.854555 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:36 crc kubenswrapper[4813]: E0317 09:11:36.954740 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:37 crc kubenswrapper[4813]: E0317 09:11:37.055760 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:37 crc kubenswrapper[4813]: E0317 09:11:37.156845 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:37 crc kubenswrapper[4813]: E0317 09:11:37.257739 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:37 crc kubenswrapper[4813]: E0317 09:11:37.358445 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:37 crc kubenswrapper[4813]: E0317 09:11:37.459099 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:37 crc kubenswrapper[4813]: E0317 09:11:37.560195 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:37 crc kubenswrapper[4813]: E0317 09:11:37.661289 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:37 crc kubenswrapper[4813]: E0317 09:11:37.761466 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:37 crc kubenswrapper[4813]: E0317 09:11:37.862425 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:37 crc kubenswrapper[4813]: E0317 09:11:37.888809 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 17 09:11:37 crc kubenswrapper[4813]: I0317 09:11:37.893882 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:37 crc kubenswrapper[4813]: I0317 09:11:37.893930 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:37 crc kubenswrapper[4813]: I0317 09:11:37.893942 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:37 crc kubenswrapper[4813]: I0317 09:11:37.893961 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:37 crc kubenswrapper[4813]: I0317 09:11:37.893974 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:37Z","lastTransitionTime":"2026-03-17T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:37 crc kubenswrapper[4813]: E0317 09:11:37.905352 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:37 crc kubenswrapper[4813]: I0317 09:11:37.910029 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:37 crc kubenswrapper[4813]: I0317 09:11:37.910100 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:37 crc kubenswrapper[4813]: I0317 09:11:37.910125 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:37 crc kubenswrapper[4813]: I0317 09:11:37.910157 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:37 crc kubenswrapper[4813]: I0317 09:11:37.910181 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:37Z","lastTransitionTime":"2026-03-17T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:37 crc kubenswrapper[4813]: E0317 09:11:37.924274 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:37 crc kubenswrapper[4813]: I0317 09:11:37.929266 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:37 crc kubenswrapper[4813]: I0317 09:11:37.929356 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:37 crc kubenswrapper[4813]: I0317 09:11:37.929368 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:37 crc kubenswrapper[4813]: I0317 09:11:37.929391 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:37 crc kubenswrapper[4813]: I0317 09:11:37.929407 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:37Z","lastTransitionTime":"2026-03-17T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:37 crc kubenswrapper[4813]: E0317 09:11:37.942307 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:37 crc kubenswrapper[4813]: I0317 09:11:37.947499 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:37 crc kubenswrapper[4813]: I0317 09:11:37.947535 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:37 crc kubenswrapper[4813]: I0317 09:11:37.947546 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:37 crc kubenswrapper[4813]: I0317 09:11:37.947562 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:37 crc kubenswrapper[4813]: I0317 09:11:37.947575 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:37Z","lastTransitionTime":"2026-03-17T09:11:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:37 crc kubenswrapper[4813]: E0317 09:11:37.962488 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:37Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:37 crc kubenswrapper[4813]: E0317 09:11:37.962648 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 17 09:11:37 crc kubenswrapper[4813]: E0317 09:11:37.962705 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:38 crc kubenswrapper[4813]: E0317 09:11:38.063589 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:38 crc kubenswrapper[4813]: E0317 09:11:38.163916 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:38 crc kubenswrapper[4813]: E0317 09:11:38.264889 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:38 crc kubenswrapper[4813]: E0317 09:11:38.365334 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:38 crc kubenswrapper[4813]: E0317 09:11:38.465450 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:38 crc kubenswrapper[4813]: E0317 09:11:38.565556 4813 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 17 09:11:38 crc kubenswrapper[4813]: E0317 09:11:38.665749 4813 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Mar 17 09:11:38 crc kubenswrapper[4813]: E0317 09:11:38.827372 4813 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 17 09:11:38 crc kubenswrapper[4813]: E0317 09:11:38.847890 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:11:41 crc kubenswrapper[4813]: I0317 09:11:41.730095 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:11:41 crc kubenswrapper[4813]: I0317 09:11:41.731583 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:41 crc kubenswrapper[4813]: I0317 09:11:41.731691 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:41 crc kubenswrapper[4813]: I0317 09:11:41.731713 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:41 crc kubenswrapper[4813]: I0317 09:11:41.732815 4813 scope.go:117] "RemoveContainer" containerID="a14d9e3f094ad9e9258a200e9020e7a49b82e421c19f2de3626fa32ba21a44e5" Mar 17 09:11:42 crc kubenswrapper[4813]: I0317 09:11:42.208116 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 17 09:11:42 crc kubenswrapper[4813]: I0317 09:11:42.209936 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f"} Mar 17 09:11:42 crc kubenswrapper[4813]: I0317 09:11:42.210047 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:11:42 crc kubenswrapper[4813]: I0317 09:11:42.210882 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:42 crc kubenswrapper[4813]: I0317 09:11:42.210947 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:42 crc kubenswrapper[4813]: I0317 09:11:42.210965 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.215455 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/4.log" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.216087 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.218648 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f" exitCode=255 Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.218663 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f"} Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.218760 4813 scope.go:117] "RemoveContainer" containerID="a14d9e3f094ad9e9258a200e9020e7a49b82e421c19f2de3626fa32ba21a44e5" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.218960 4813 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.220289 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.220341 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.220353 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.221309 4813 scope.go:117] "RemoveContainer" containerID="05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f" Mar 17 09:11:43 crc kubenswrapper[4813]: E0317 09:11:43.221513 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.458036 4813 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.690577 4813 apiserver.go:52] "Watching apiserver" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.697364 4813 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.698156 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-7xxkf","openshift-multus/multus-5jjhb","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-multus/multus-additional-cni-plugins-c2dwc","openshift-network-node-identity/network-node-identity-vrzqb","openshift-ovn-kubernetes/ovnkube-node-lngs6","openshift-image-registry/node-ca-62stj","openshift-multus/network-metrics-daemon-l47ql","openshift-network-diagnostics/network-check-target-xd92c","openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp","openshift-dns/node-resolver-zr9rn","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.698717 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:11:43 crc kubenswrapper[4813]: E0317 09:11:43.698805 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.698877 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.699434 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.699516 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:11:43 crc kubenswrapper[4813]: E0317 09:11:43.699688 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.699778 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.699881 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 17 09:11:43 crc kubenswrapper[4813]: E0317 09:11:43.700824 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.701732 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.702397 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.703064 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-62stj" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.707126 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.707283 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:11:43 crc kubenswrapper[4813]: E0317 09:11:43.707624 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.708195 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.708363 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-zr9rn" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.709749 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.710121 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.710132 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.710128 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.713649 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.713870 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.714019 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.714200 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.714486 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.714806 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.714924 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.714963 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.714989 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.715170 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.715233 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.715347 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.716015 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.716065 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.715761 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.716348 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.716348 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.716564 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.716668 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.717012 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.717034 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.717130 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.717196 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.717355 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.717393 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.717495 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.717709 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.717844 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.726946 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.727191 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.727285 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.727692 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.727845 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.728015 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.735731 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.754731 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.767299 4813 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.771478 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.787210 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.801135 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.801219 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.801273 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.801321 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.801405 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.801450 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.801550 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.801640 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.801699 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.801744 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.801786 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.801830 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.801879 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.801927 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.801972 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.802020 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.802063 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.802111 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.802154 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.802200 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.802250 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.802415 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.802470 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.802517 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.802563 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.802646 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.802695 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.802743 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.802794 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.802841 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.802893 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.802940 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.802991 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803039 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803092 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803085 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803144 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803193 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803240 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803288 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803334 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803386 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803434 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803481 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803525 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803504 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803529 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803574 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803669 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803720 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803727 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803768 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803820 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803875 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803932 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.803976 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.804024 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.804074 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.804124 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.804131 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.804164 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.804176 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.804240 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.804300 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.804348 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.804394 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 17 09:11:43 crc kubenswrapper[4813]: E0317 09:11:43.804434 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:11:44.30439489 +0000 UTC m=+126.405198419 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.804460 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.804496 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.804698 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.804745 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.804784 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.804804 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.804822 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.804920 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.804958 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.804989 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.805035 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.805129 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.805255 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.805306 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.805388 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.805520 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.805576 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.805664 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.805710 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.805760 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.805807 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.805851 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.805911 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.805958 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806006 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806057 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806105 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806150 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806198 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806250 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806306 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806355 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806407 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806454 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806506 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806552 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806635 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806689 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806741 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806791 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806837 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806885 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806935 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806981 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807028 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807075 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807131 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807183 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807232 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807275 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807320 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807363 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807414 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807464 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807534 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807586 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807670 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807721 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807771 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807817 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807868 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807921 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807974 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808024 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808077 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808127 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808175 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808219 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808261 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808314 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808364 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808412 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808461 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808519 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808568 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808649 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808704 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808750 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808799 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808850 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808896 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808945 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808995 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809043 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809098 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809148 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809205 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809257 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809313 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809405 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809472 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809529 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809578 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809719 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809783 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809840 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809889 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809946 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809997 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.810052 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.810106 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.810159 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.810214 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.810275 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.810329 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.810378 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.810432 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.810509 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.810559 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.810642 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.810700 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.811209 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.811276 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.811329 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.811384 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.811436 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.811490 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.811544 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.811836 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.811908 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.811946 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.811975 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812007 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812035 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812073 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812108 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812135 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812240 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812281 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812309 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812341 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812366 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812396 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812423 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812452 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812478 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812506 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812535 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812562 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812689 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812730 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs\") pod \"network-metrics-daemon-l47ql\" (UID: \"27d4c184-5dd3-492a-b927-f7a7f31291ef\") " pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812756 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wd8rt\" (UniqueName: \"kubernetes.io/projected/27d4c184-5dd3-492a-b927-f7a7f31291ef-kube-api-access-wd8rt\") pod \"network-metrics-daemon-l47ql\" (UID: \"27d4c184-5dd3-492a-b927-f7a7f31291ef\") " pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812797 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812829 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812883 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-kubelet\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812911 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-node-log\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812942 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d35d2ade-2829-4fc2-864d-1871c18006f7-tuning-conf-dir\") pod \"multus-additional-cni-plugins-c2dwc\" (UID: \"d35d2ade-2829-4fc2-864d-1871c18006f7\") " pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812973 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812999 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8ac387df-f9b1-4ce9-a109-1b80e7659f2e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bb4gp\" (UID: \"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813030 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-multus-conf-dir\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813058 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d35d2ade-2829-4fc2-864d-1871c18006f7-system-cni-dir\") pod \"multus-additional-cni-plugins-c2dwc\" (UID: \"d35d2ade-2829-4fc2-864d-1871c18006f7\") " pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813084 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e7e156ce-b279-4a25-a27e-2da7d0ba6437-rootfs\") pod \"machine-config-daemon-7xxkf\" (UID: \"e7e156ce-b279-4a25-a27e-2da7d0ba6437\") " pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813111 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-env-overrides\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813141 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zs4ft\" (UniqueName: \"kubernetes.io/projected/8a42f832-caac-4d43-aba1-63eab74eb5f5-kube-api-access-zs4ft\") pod \"node-ca-62stj\" (UID: \"8a42f832-caac-4d43-aba1-63eab74eb5f5\") " pod="openshift-image-registry/node-ca-62stj" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813166 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8ac387df-f9b1-4ce9-a109-1b80e7659f2e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bb4gp\" (UID: \"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813191 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-multus-socket-dir-parent\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813219 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-systemd-units\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813247 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m6hct\" (UniqueName: \"kubernetes.io/projected/95fc4db5-3177-485a-adfd-f829495e487a-kube-api-access-m6hct\") pod \"node-resolver-zr9rn\" (UID: \"95fc4db5-3177-485a-adfd-f829495e487a\") " pod="openshift-dns/node-resolver-zr9rn" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813273 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-system-cni-dir\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813295 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-etc-openvswitch\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813323 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-host-var-lib-cni-bin\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813351 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-run-systemd\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813468 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-ovnkube-config\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813496 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e7e156ce-b279-4a25-a27e-2da7d0ba6437-proxy-tls\") pod \"machine-config-daemon-7xxkf\" (UID: \"e7e156ce-b279-4a25-a27e-2da7d0ba6437\") " pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813521 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-run-openvswitch\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813551 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813577 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-cnibin\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813698 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d35d2ade-2829-4fc2-864d-1871c18006f7-cni-binary-copy\") pod \"multus-additional-cni-plugins-c2dwc\" (UID: \"d35d2ade-2829-4fc2-864d-1871c18006f7\") " pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813729 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j4jw5\" (UniqueName: \"kubernetes.io/projected/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-kube-api-access-j4jw5\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813753 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8a42f832-caac-4d43-aba1-63eab74eb5f5-serviceca\") pod \"node-ca-62stj\" (UID: \"8a42f832-caac-4d43-aba1-63eab74eb5f5\") " pod="openshift-image-registry/node-ca-62stj" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813777 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-26dhg\" (UniqueName: \"kubernetes.io/projected/e7e156ce-b279-4a25-a27e-2da7d0ba6437-kube-api-access-26dhg\") pod \"machine-config-daemon-7xxkf\" (UID: \"e7e156ce-b279-4a25-a27e-2da7d0ba6437\") " pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813803 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-multus-cni-dir\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813829 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-var-lib-openvswitch\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813858 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-ovn-node-metrics-cert\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813880 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8a42f832-caac-4d43-aba1-63eab74eb5f5-host\") pod \"node-ca-62stj\" (UID: \"8a42f832-caac-4d43-aba1-63eab74eb5f5\") " pod="openshift-image-registry/node-ca-62stj" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813902 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813925 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/95fc4db5-3177-485a-adfd-f829495e487a-hosts-file\") pod \"node-resolver-zr9rn\" (UID: \"95fc4db5-3177-485a-adfd-f829495e487a\") " pod="openshift-dns/node-resolver-zr9rn" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813955 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813987 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814014 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-run-ovn-kubernetes\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814039 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-run-ovn\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814064 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-log-socket\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814087 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-cni-bin\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814111 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-cni-netd\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814138 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-run-netns\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814188 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814213 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-host-run-netns\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814237 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d35d2ade-2829-4fc2-864d-1871c18006f7-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-c2dwc\" (UID: \"d35d2ade-2829-4fc2-864d-1871c18006f7\") " pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814261 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmxwh\" (UniqueName: \"kubernetes.io/projected/d35d2ade-2829-4fc2-864d-1871c18006f7-kube-api-access-wmxwh\") pod \"multus-additional-cni-plugins-c2dwc\" (UID: \"d35d2ade-2829-4fc2-864d-1871c18006f7\") " pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814294 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814320 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-host-run-k8s-cni-cncf-io\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814344 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-host-var-lib-kubelet\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814367 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d35d2ade-2829-4fc2-864d-1871c18006f7-cnibin\") pod \"multus-additional-cni-plugins-c2dwc\" (UID: \"d35d2ade-2829-4fc2-864d-1871c18006f7\") " pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814393 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814416 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-host-var-lib-cni-multus\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814438 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c1246d4d-93d5-4a97-bef8-1ed881e1a217-multus-daemon-config\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814464 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-etc-kubernetes\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814489 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d35d2ade-2829-4fc2-864d-1871c18006f7-os-release\") pod \"multus-additional-cni-plugins-c2dwc\" (UID: \"d35d2ade-2829-4fc2-864d-1871c18006f7\") " pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814520 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814547 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814574 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e7e156ce-b279-4a25-a27e-2da7d0ba6437-mcd-auth-proxy-config\") pod \"machine-config-daemon-7xxkf\" (UID: \"e7e156ce-b279-4a25-a27e-2da7d0ba6437\") " pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814636 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-os-release\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814677 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-host-run-multus-certs\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814710 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814743 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814771 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvftk\" (UniqueName: \"kubernetes.io/projected/8ac387df-f9b1-4ce9-a109-1b80e7659f2e-kube-api-access-xvftk\") pod \"ovnkube-control-plane-749d76644c-bb4gp\" (UID: \"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814798 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c1246d4d-93d5-4a97-bef8-1ed881e1a217-cni-binary-copy\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814821 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-hostroot\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814848 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mpt6d\" (UniqueName: \"kubernetes.io/projected/c1246d4d-93d5-4a97-bef8-1ed881e1a217-kube-api-access-mpt6d\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814874 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-slash\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814899 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-ovnkube-script-lib\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814930 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8ac387df-f9b1-4ce9-a109-1b80e7659f2e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bb4gp\" (UID: \"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.815077 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.815097 4813 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.815119 4813 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.815139 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.815157 4813 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.815173 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.815191 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.815208 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.815223 4813 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.815239 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.805444 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.805528 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806063 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806678 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.826539 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.826579 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.806717 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807096 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807101 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807259 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807437 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807707 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.807859 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808472 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808533 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808620 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.808974 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809191 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809164 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809297 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809479 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809772 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809977 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.809968 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.810104 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.810212 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.810863 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.810945 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.811257 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.811852 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812106 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812386 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812454 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812514 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812712 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812845 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.812898 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813053 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813096 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813160 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813540 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813782 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814575 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814614 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814631 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814720 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814873 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.814922 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.815646 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.816055 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.816081 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.816092 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.816574 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.816788 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.816939 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.817789 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.818202 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.818264 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.818448 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.818551 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.818671 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.818990 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.819468 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.820445 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.820822 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.821198 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.821235 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.821263 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.821280 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.821721 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.822028 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.822062 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.822312 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.822393 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.822860 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.822886 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.823136 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.823591 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.823671 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.823687 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.823746 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.823893 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.823791 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.824005 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.824176 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.825086 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.825192 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.825211 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.826065 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.826105 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.826315 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.825915 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.832855 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.833314 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.833363 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.833562 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.833949 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.834865 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: E0317 09:11:43.835090 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 09:11:43 crc kubenswrapper[4813]: E0317 09:11:43.835260 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 09:11:44.33522988 +0000 UTC m=+126.436033389 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.836730 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.837731 4813 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.838791 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.840624 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.841560 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.842131 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.842514 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.843230 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.843550 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.813169 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.843811 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.844195 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.844230 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.845250 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.846200 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.847985 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.849313 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.852031 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.854319 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: E0317 09:11:43.855141 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 09:11:43 crc kubenswrapper[4813]: E0317 09:11:43.855173 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 09:11:43 crc kubenswrapper[4813]: E0317 09:11:43.855188 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:43 crc kubenswrapper[4813]: E0317 09:11:43.855258 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-17 09:11:44.355236262 +0000 UTC m=+126.456039761 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.858297 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.859277 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.860906 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.859421 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.861489 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 17 09:11:43 crc kubenswrapper[4813]: E0317 09:11:43.861957 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:11:43 crc kubenswrapper[4813]: E0317 09:11:43.859564 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.862145 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.861217 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.859679 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.860510 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.860621 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.861186 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.861819 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.862035 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.862163 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: E0317 09:11:43.862107 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.862742 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: E0317 09:11:43.863489 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 09:11:44.363397892 +0000 UTC m=+126.464201401 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 09:11:43 crc kubenswrapper[4813]: E0317 09:11:43.865066 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 09:11:43 crc kubenswrapper[4813]: E0317 09:11:43.865105 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:43 crc kubenswrapper[4813]: E0317 09:11:43.865187 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-17 09:11:44.36516452 +0000 UTC m=+126.465968029 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.866981 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.868093 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.868172 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.868544 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.868971 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.869196 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.869384 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.869501 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.869524 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.869659 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.869765 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.869768 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.870035 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.870472 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.870908 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.871675 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.872069 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.872266 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.872758 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.872905 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.872971 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.872960 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.873121 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.873169 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.873507 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.873530 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.873694 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.873740 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.874110 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.874198 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.874357 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.875420 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.875404 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.876113 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.876572 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.876715 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.876850 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.878525 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.878693 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.878721 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.879217 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.879413 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.879500 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.879557 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.879572 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.880178 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.880568 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.880732 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.882659 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.884424 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.884541 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.885190 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.885495 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.885577 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.885702 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.885747 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.885757 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.885708 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.885831 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.886875 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.888049 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.893172 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.893304 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.899657 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.906409 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.914525 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916033 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-system-cni-dir\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916083 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-multus-socket-dir-parent\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916165 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-system-cni-dir\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916276 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-systemd-units\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916315 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m6hct\" (UniqueName: \"kubernetes.io/projected/95fc4db5-3177-485a-adfd-f829495e487a-kube-api-access-m6hct\") pod \"node-resolver-zr9rn\" (UID: \"95fc4db5-3177-485a-adfd-f829495e487a\") " pod="openshift-dns/node-resolver-zr9rn" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916337 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-etc-openvswitch\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916357 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e7e156ce-b279-4a25-a27e-2da7d0ba6437-proxy-tls\") pod \"machine-config-daemon-7xxkf\" (UID: \"e7e156ce-b279-4a25-a27e-2da7d0ba6437\") " pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916363 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-multus-socket-dir-parent\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916381 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-systemd-units\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916423 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-host-var-lib-cni-bin\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916421 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-etc-openvswitch\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916374 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-host-var-lib-cni-bin\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916517 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-run-systemd\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916548 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-ovnkube-config\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916573 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-run-openvswitch\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916622 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-multus-cni-dir\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916648 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-cnibin\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916672 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d35d2ade-2829-4fc2-864d-1871c18006f7-cni-binary-copy\") pod \"multus-additional-cni-plugins-c2dwc\" (UID: \"d35d2ade-2829-4fc2-864d-1871c18006f7\") " pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916705 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j4jw5\" (UniqueName: \"kubernetes.io/projected/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-kube-api-access-j4jw5\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916729 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8a42f832-caac-4d43-aba1-63eab74eb5f5-serviceca\") pod \"node-ca-62stj\" (UID: \"8a42f832-caac-4d43-aba1-63eab74eb5f5\") " pod="openshift-image-registry/node-ca-62stj" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916751 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-26dhg\" (UniqueName: \"kubernetes.io/projected/e7e156ce-b279-4a25-a27e-2da7d0ba6437-kube-api-access-26dhg\") pod \"machine-config-daemon-7xxkf\" (UID: \"e7e156ce-b279-4a25-a27e-2da7d0ba6437\") " pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916777 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8a42f832-caac-4d43-aba1-63eab74eb5f5-host\") pod \"node-ca-62stj\" (UID: \"8a42f832-caac-4d43-aba1-63eab74eb5f5\") " pod="openshift-image-registry/node-ca-62stj" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916775 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-run-systemd\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916827 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-var-lib-openvswitch\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916801 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-var-lib-openvswitch\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916925 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-run-openvswitch\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916929 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-ovn-node-metrics-cert\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.916980 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-run-ovn-kubernetes\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917002 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917022 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/95fc4db5-3177-485a-adfd-f829495e487a-hosts-file\") pod \"node-resolver-zr9rn\" (UID: \"95fc4db5-3177-485a-adfd-f829495e487a\") " pod="openshift-dns/node-resolver-zr9rn" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917070 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-run-netns\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917090 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-run-ovn\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917108 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-log-socket\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917125 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-cni-bin\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917142 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-cni-netd\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917139 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-cnibin\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917154 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-multus-cni-dir\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917166 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmxwh\" (UniqueName: \"kubernetes.io/projected/d35d2ade-2829-4fc2-864d-1871c18006f7-kube-api-access-wmxwh\") pod \"multus-additional-cni-plugins-c2dwc\" (UID: \"d35d2ade-2829-4fc2-864d-1871c18006f7\") " pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917199 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-run-ovn-kubernetes\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917206 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-run-ovn\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917228 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-log-socket\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917262 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-cni-bin\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917270 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-run-netns\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917284 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-cni-netd\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917328 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917446 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/95fc4db5-3177-485a-adfd-f829495e487a-hosts-file\") pod \"node-resolver-zr9rn\" (UID: \"95fc4db5-3177-485a-adfd-f829495e487a\") " pod="openshift-dns/node-resolver-zr9rn" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917552 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8a42f832-caac-4d43-aba1-63eab74eb5f5-host\") pod \"node-ca-62stj\" (UID: \"8a42f832-caac-4d43-aba1-63eab74eb5f5\") " pod="openshift-image-registry/node-ca-62stj" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917622 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-host-run-netns\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917660 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d35d2ade-2829-4fc2-864d-1871c18006f7-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-c2dwc\" (UID: \"d35d2ade-2829-4fc2-864d-1871c18006f7\") " pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917687 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-host-run-k8s-cni-cncf-io\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917713 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-host-var-lib-kubelet\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917787 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-etc-kubernetes\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917813 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d35d2ade-2829-4fc2-864d-1871c18006f7-cnibin\") pod \"multus-additional-cni-plugins-c2dwc\" (UID: \"d35d2ade-2829-4fc2-864d-1871c18006f7\") " pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917836 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-host-var-lib-cni-multus\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917858 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c1246d4d-93d5-4a97-bef8-1ed881e1a217-multus-daemon-config\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917880 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d35d2ade-2829-4fc2-864d-1871c18006f7-os-release\") pod \"multus-additional-cni-plugins-c2dwc\" (UID: \"d35d2ade-2829-4fc2-864d-1871c18006f7\") " pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917909 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e7e156ce-b279-4a25-a27e-2da7d0ba6437-mcd-auth-proxy-config\") pod \"machine-config-daemon-7xxkf\" (UID: \"e7e156ce-b279-4a25-a27e-2da7d0ba6437\") " pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917929 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-os-release\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917964 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-host-run-multus-certs\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.917989 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918017 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8ac387df-f9b1-4ce9-a109-1b80e7659f2e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bb4gp\" (UID: \"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918048 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvftk\" (UniqueName: \"kubernetes.io/projected/8ac387df-f9b1-4ce9-a109-1b80e7659f2e-kube-api-access-xvftk\") pod \"ovnkube-control-plane-749d76644c-bb4gp\" (UID: \"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918078 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c1246d4d-93d5-4a97-bef8-1ed881e1a217-cni-binary-copy\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918100 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-hostroot\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918119 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mpt6d\" (UniqueName: \"kubernetes.io/projected/c1246d4d-93d5-4a97-bef8-1ed881e1a217-kube-api-access-mpt6d\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918142 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-slash\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918163 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-ovnkube-script-lib\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918183 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-node-log\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918203 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918226 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs\") pod \"network-metrics-daemon-l47ql\" (UID: \"27d4c184-5dd3-492a-b927-f7a7f31291ef\") " pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918247 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wd8rt\" (UniqueName: \"kubernetes.io/projected/27d4c184-5dd3-492a-b927-f7a7f31291ef-kube-api-access-wd8rt\") pod \"network-metrics-daemon-l47ql\" (UID: \"27d4c184-5dd3-492a-b927-f7a7f31291ef\") " pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918280 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-kubelet\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918304 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d35d2ade-2829-4fc2-864d-1871c18006f7-system-cni-dir\") pod \"multus-additional-cni-plugins-c2dwc\" (UID: \"d35d2ade-2829-4fc2-864d-1871c18006f7\") " pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918328 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d35d2ade-2829-4fc2-864d-1871c18006f7-tuning-conf-dir\") pod \"multus-additional-cni-plugins-c2dwc\" (UID: \"d35d2ade-2829-4fc2-864d-1871c18006f7\") " pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918357 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8ac387df-f9b1-4ce9-a109-1b80e7659f2e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bb4gp\" (UID: \"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918386 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-multus-conf-dir\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918418 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8ac387df-f9b1-4ce9-a109-1b80e7659f2e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bb4gp\" (UID: \"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918438 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d35d2ade-2829-4fc2-864d-1871c18006f7-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-c2dwc\" (UID: \"d35d2ade-2829-4fc2-864d-1871c18006f7\") " pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918440 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e7e156ce-b279-4a25-a27e-2da7d0ba6437-rootfs\") pod \"machine-config-daemon-7xxkf\" (UID: \"e7e156ce-b279-4a25-a27e-2da7d0ba6437\") " pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918469 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/e7e156ce-b279-4a25-a27e-2da7d0ba6437-rootfs\") pod \"machine-config-daemon-7xxkf\" (UID: \"e7e156ce-b279-4a25-a27e-2da7d0ba6437\") " pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918489 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-env-overrides\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918499 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-host-run-netns\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918511 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zs4ft\" (UniqueName: \"kubernetes.io/projected/8a42f832-caac-4d43-aba1-63eab74eb5f5-kube-api-access-zs4ft\") pod \"node-ca-62stj\" (UID: \"8a42f832-caac-4d43-aba1-63eab74eb5f5\") " pod="openshift-image-registry/node-ca-62stj" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918552 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d35d2ade-2829-4fc2-864d-1871c18006f7-os-release\") pod \"multus-additional-cni-plugins-c2dwc\" (UID: \"d35d2ade-2829-4fc2-864d-1871c18006f7\") " pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918633 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d35d2ade-2829-4fc2-864d-1871c18006f7-cnibin\") pod \"multus-additional-cni-plugins-c2dwc\" (UID: \"d35d2ade-2829-4fc2-864d-1871c18006f7\") " pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918662 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-host-var-lib-cni-multus\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918512 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-hostroot\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918469 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8a42f832-caac-4d43-aba1-63eab74eb5f5-serviceca\") pod \"node-ca-62stj\" (UID: \"8a42f832-caac-4d43-aba1-63eab74eb5f5\") " pod="openshift-image-registry/node-ca-62stj" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.918992 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-env-overrides\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.919300 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.919335 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-multus-conf-dir\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.919364 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-slash\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.919885 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-ovnkube-script-lib\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.919912 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-node-log\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.919931 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: E0317 09:11:43.919991 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 09:11:43 crc kubenswrapper[4813]: E0317 09:11:43.920038 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs podName:27d4c184-5dd3-492a-b927-f7a7f31291ef nodeName:}" failed. No retries permitted until 2026-03-17 09:11:44.420026315 +0000 UTC m=+126.520829804 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs") pod "network-metrics-daemon-l47ql" (UID: "27d4c184-5dd3-492a-b927-f7a7f31291ef") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.920186 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/c1246d4d-93d5-4a97-bef8-1ed881e1a217-multus-daemon-config\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.920223 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-etc-kubernetes\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.920274 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-kubelet\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.920298 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d35d2ade-2829-4fc2-864d-1871c18006f7-system-cni-dir\") pod \"multus-additional-cni-plugins-c2dwc\" (UID: \"d35d2ade-2829-4fc2-864d-1871c18006f7\") " pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.920411 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-host-var-lib-kubelet\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.920439 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-host-run-k8s-cni-cncf-io\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.920455 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-host-run-multus-certs\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.920567 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.921009 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d35d2ade-2829-4fc2-864d-1871c18006f7-tuning-conf-dir\") pod \"multus-additional-cni-plugins-c2dwc\" (UID: \"d35d2ade-2829-4fc2-864d-1871c18006f7\") " pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.921583 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/e7e156ce-b279-4a25-a27e-2da7d0ba6437-mcd-auth-proxy-config\") pod \"machine-config-daemon-7xxkf\" (UID: \"e7e156ce-b279-4a25-a27e-2da7d0ba6437\") " pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.921613 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/c1246d4d-93d5-4a97-bef8-1ed881e1a217-cni-binary-copy\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.921647 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.921698 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/c1246d4d-93d5-4a97-bef8-1ed881e1a217-os-release\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.921857 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.921871 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.921881 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.921892 4813 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.921902 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.921911 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.921921 4813 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.921930 4813 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.921940 4813 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.921948 4813 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.921962 4813 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.921972 4813 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.921981 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.921991 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922000 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922009 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922018 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922028 4813 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922037 4813 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922047 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922056 4813 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922065 4813 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922075 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922083 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922093 4813 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922103 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922112 4813 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922122 4813 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922132 4813 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922141 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922150 4813 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922159 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922168 4813 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922177 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922187 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922196 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922205 4813 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922214 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922223 4813 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922233 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922242 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922250 4813 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922260 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922270 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922280 4813 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922288 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922296 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922305 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922314 4813 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922322 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922330 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922338 4813 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922347 4813 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922355 4813 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922364 4813 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922373 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922386 4813 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922397 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922406 4813 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922415 4813 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922424 4813 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922433 4813 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922442 4813 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922453 4813 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922463 4813 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922473 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922482 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922491 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922501 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922624 4813 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922637 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922646 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922664 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922680 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922692 4813 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922705 4813 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.919733 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922717 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922681 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8ac387df-f9b1-4ce9-a109-1b80e7659f2e-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-bb4gp\" (UID: \"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922728 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922736 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922746 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922756 4813 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922767 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922777 4813 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922785 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922794 4813 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922803 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922811 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922822 4813 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922831 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922840 4813 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922850 4813 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922858 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922867 4813 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922876 4813 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922884 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922893 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922902 4813 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922912 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922921 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922929 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922938 4813 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922948 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922956 4813 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922966 4813 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922975 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922984 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.922992 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923003 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923011 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923020 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923029 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923037 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923046 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923054 4813 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923063 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923071 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923086 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923094 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923103 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923111 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923119 4813 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923129 4813 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923137 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923147 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923155 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923163 4813 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923171 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923179 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923188 4813 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923198 4813 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923208 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923217 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923226 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923234 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923242 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923251 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923260 4813 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923269 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923277 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923286 4813 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923294 4813 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923302 4813 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923311 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923320 4813 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923329 4813 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923337 4813 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923345 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923353 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923362 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923370 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923379 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923388 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923396 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923406 4813 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923416 4813 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923424 4813 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923433 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923441 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923451 4813 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923459 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923468 4813 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923475 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923483 4813 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923491 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923499 4813 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923510 4813 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923519 4813 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923531 4813 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923539 4813 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923547 4813 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923558 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923568 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923577 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923584 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923592 4813 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923632 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923641 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923650 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923659 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923667 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923675 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.923683 4813 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.931815 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-ovnkube-config\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.932341 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.932966 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d35d2ade-2829-4fc2-864d-1871c18006f7-cni-binary-copy\") pod \"multus-additional-cni-plugins-c2dwc\" (UID: \"d35d2ade-2829-4fc2-864d-1871c18006f7\") " pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.933317 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8ac387df-f9b1-4ce9-a109-1b80e7659f2e-env-overrides\") pod \"ovnkube-control-plane-749d76644c-bb4gp\" (UID: \"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.934422 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8ac387df-f9b1-4ce9-a109-1b80e7659f2e-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-bb4gp\" (UID: \"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.934751 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-ovn-node-metrics-cert\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.936893 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmxwh\" (UniqueName: \"kubernetes.io/projected/d35d2ade-2829-4fc2-864d-1871c18006f7-kube-api-access-wmxwh\") pod \"multus-additional-cni-plugins-c2dwc\" (UID: \"d35d2ade-2829-4fc2-864d-1871c18006f7\") " pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.936968 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j4jw5\" (UniqueName: \"kubernetes.io/projected/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-kube-api-access-j4jw5\") pod \"ovnkube-node-lngs6\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.937734 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvftk\" (UniqueName: \"kubernetes.io/projected/8ac387df-f9b1-4ce9-a109-1b80e7659f2e-kube-api-access-xvftk\") pod \"ovnkube-control-plane-749d76644c-bb4gp\" (UID: \"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.938844 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/e7e156ce-b279-4a25-a27e-2da7d0ba6437-proxy-tls\") pod \"machine-config-daemon-7xxkf\" (UID: \"e7e156ce-b279-4a25-a27e-2da7d0ba6437\") " pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.940087 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m6hct\" (UniqueName: \"kubernetes.io/projected/95fc4db5-3177-485a-adfd-f829495e487a-kube-api-access-m6hct\") pod \"node-resolver-zr9rn\" (UID: \"95fc4db5-3177-485a-adfd-f829495e487a\") " pod="openshift-dns/node-resolver-zr9rn" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.940260 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-26dhg\" (UniqueName: \"kubernetes.io/projected/e7e156ce-b279-4a25-a27e-2da7d0ba6437-kube-api-access-26dhg\") pod \"machine-config-daemon-7xxkf\" (UID: \"e7e156ce-b279-4a25-a27e-2da7d0ba6437\") " pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.940432 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.942067 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wd8rt\" (UniqueName: \"kubernetes.io/projected/27d4c184-5dd3-492a-b927-f7a7f31291ef-kube-api-access-wd8rt\") pod \"network-metrics-daemon-l47ql\" (UID: \"27d4c184-5dd3-492a-b927-f7a7f31291ef\") " pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.946058 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mpt6d\" (UniqueName: \"kubernetes.io/projected/c1246d4d-93d5-4a97-bef8-1ed881e1a217-kube-api-access-mpt6d\") pod \"multus-5jjhb\" (UID: \"c1246d4d-93d5-4a97-bef8-1ed881e1a217\") " pod="openshift-multus/multus-5jjhb" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.947051 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zs4ft\" (UniqueName: \"kubernetes.io/projected/8a42f832-caac-4d43-aba1-63eab74eb5f5-kube-api-access-zs4ft\") pod \"node-ca-62stj\" (UID: \"8a42f832-caac-4d43-aba1-63eab74eb5f5\") " pod="openshift-image-registry/node-ca-62stj" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.949559 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.958257 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:43 crc kubenswrapper[4813]: I0317 09:11:43.966208 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.019742 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 17 09:11:44 crc kubenswrapper[4813]: W0317 09:11:44.037628 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-53b7075e2a2f13919ee3688f52d64128acba653a18357a5fb496b951e2c966cf WatchSource:0}: Error finding container 53b7075e2a2f13919ee3688f52d64128acba653a18357a5fb496b951e2c966cf: Status 404 returned error can't find the container with id 53b7075e2a2f13919ee3688f52d64128acba653a18357a5fb496b951e2c966cf Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.043059 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.060372 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 17 09:11:44 crc kubenswrapper[4813]: W0317 09:11:44.060910 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-7c62a43f8c342b0044cd35d5bd17c198a9eafc72cb8e7cd3dacc24dcb74693be WatchSource:0}: Error finding container 7c62a43f8c342b0044cd35d5bd17c198a9eafc72cb8e7cd3dacc24dcb74693be: Status 404 returned error can't find the container with id 7c62a43f8c342b0044cd35d5bd17c198a9eafc72cb8e7cd3dacc24dcb74693be Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.077145 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:44 crc kubenswrapper[4813]: W0317 09:11:44.078353 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-62c289c4ee90a35c13954bcbb213e74908b50923fe97a3e3241c7d89a7aadc01 WatchSource:0}: Error finding container 62c289c4ee90a35c13954bcbb213e74908b50923fe97a3e3241c7d89a7aadc01: Status 404 returned error can't find the container with id 62c289c4ee90a35c13954bcbb213e74908b50923fe97a3e3241c7d89a7aadc01 Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.096448 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.112929 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:11:44 crc kubenswrapper[4813]: W0317 09:11:44.128964 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd35d2ade_2829_4fc2_864d_1871c18006f7.slice/crio-f15b5bffa734cc29e46442b8ce1d0ef8d51dc831b07759c4380c31566dbcc8eb WatchSource:0}: Error finding container f15b5bffa734cc29e46442b8ce1d0ef8d51dc831b07759c4380c31566dbcc8eb: Status 404 returned error can't find the container with id f15b5bffa734cc29e46442b8ce1d0ef8d51dc831b07759c4380c31566dbcc8eb Mar 17 09:11:44 crc kubenswrapper[4813]: W0317 09:11:44.138581 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode7e156ce_b279_4a25_a27e_2da7d0ba6437.slice/crio-a7b6a52b02b9b0f89a40b62a9106fc13b10230b3e754d88e7e14faa881ce272e WatchSource:0}: Error finding container a7b6a52b02b9b0f89a40b62a9106fc13b10230b3e754d88e7e14faa881ce272e: Status 404 returned error can't find the container with id a7b6a52b02b9b0f89a40b62a9106fc13b10230b3e754d88e7e14faa881ce272e Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.141592 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.148145 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-zr9rn" Mar 17 09:11:44 crc kubenswrapper[4813]: W0317 09:11:44.176113 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8ac387df_f9b1_4ce9_a109_1b80e7659f2e.slice/crio-8efa3190c477a4c63ad38439c7a194e35f3c56308936fea73e2c79333481e980 WatchSource:0}: Error finding container 8efa3190c477a4c63ad38439c7a194e35f3c56308936fea73e2c79333481e980: Status 404 returned error can't find the container with id 8efa3190c477a4c63ad38439c7a194e35f3c56308936fea73e2c79333481e980 Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.183698 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-62stj" Mar 17 09:11:44 crc kubenswrapper[4813]: W0317 09:11:44.186917 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod95fc4db5_3177_485a_adfd_f829495e487a.slice/crio-803c7f301cabf861c655a32013125a4677e7c9562f83c6a58e2a815e269e42e2 WatchSource:0}: Error finding container 803c7f301cabf861c655a32013125a4677e7c9562f83c6a58e2a815e269e42e2: Status 404 returned error can't find the container with id 803c7f301cabf861c655a32013125a4677e7c9562f83c6a58e2a815e269e42e2 Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.195713 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-5jjhb" Mar 17 09:11:44 crc kubenswrapper[4813]: W0317 09:11:44.202155 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8a42f832_caac_4d43_aba1_63eab74eb5f5.slice/crio-e72be05d57acb725a2362d2cfa77718a4d0b40edc783d542f3f815b31dff93b5 WatchSource:0}: Error finding container e72be05d57acb725a2362d2cfa77718a4d0b40edc783d542f3f815b31dff93b5: Status 404 returned error can't find the container with id e72be05d57acb725a2362d2cfa77718a4d0b40edc783d542f3f815b31dff93b5 Mar 17 09:11:44 crc kubenswrapper[4813]: W0317 09:11:44.229140 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc1246d4d_93d5_4a97_bef8_1ed881e1a217.slice/crio-efeb97f54353449254945237e34ed0adebbd28c0ad922eeea06bed59539ee5e4 WatchSource:0}: Error finding container efeb97f54353449254945237e34ed0adebbd28c0ad922eeea06bed59539ee5e4: Status 404 returned error can't find the container with id efeb97f54353449254945237e34ed0adebbd28c0ad922eeea06bed59539ee5e4 Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.239240 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/4.log" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.242924 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-62stj" event={"ID":"8a42f832-caac-4d43-aba1-63eab74eb5f5","Type":"ContainerStarted","Data":"e72be05d57acb725a2362d2cfa77718a4d0b40edc783d542f3f815b31dff93b5"} Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.247295 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" event={"ID":"8ac387df-f9b1-4ce9-a109-1b80e7659f2e","Type":"ContainerStarted","Data":"8efa3190c477a4c63ad38439c7a194e35f3c56308936fea73e2c79333481e980"} Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.249727 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"7c62a43f8c342b0044cd35d5bd17c198a9eafc72cb8e7cd3dacc24dcb74693be"} Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.251472 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"53b7075e2a2f13919ee3688f52d64128acba653a18357a5fb496b951e2c966cf"} Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.253398 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerStarted","Data":"4e31eab6db6123fdbd4bd448a5938c83e568c948e29eec8e6cb4a07dc82b2f09"} Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.254397 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerStarted","Data":"a7b6a52b02b9b0f89a40b62a9106fc13b10230b3e754d88e7e14faa881ce272e"} Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.255315 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"62c289c4ee90a35c13954bcbb213e74908b50923fe97a3e3241c7d89a7aadc01"} Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.259139 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" event={"ID":"d35d2ade-2829-4fc2-864d-1871c18006f7","Type":"ContainerStarted","Data":"f15b5bffa734cc29e46442b8ce1d0ef8d51dc831b07759c4380c31566dbcc8eb"} Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.262225 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-zr9rn" event={"ID":"95fc4db5-3177-485a-adfd-f829495e487a","Type":"ContainerStarted","Data":"803c7f301cabf861c655a32013125a4677e7c9562f83c6a58e2a815e269e42e2"} Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.334399 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:11:44 crc kubenswrapper[4813]: E0317 09:11:44.334589 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:11:45.334547815 +0000 UTC m=+127.435351314 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.435188 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.435227 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.435266 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.435304 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs\") pod \"network-metrics-daemon-l47ql\" (UID: \"27d4c184-5dd3-492a-b927-f7a7f31291ef\") " pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.435331 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:11:44 crc kubenswrapper[4813]: E0317 09:11:44.435404 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 09:11:44 crc kubenswrapper[4813]: E0317 09:11:44.435428 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 09:11:44 crc kubenswrapper[4813]: E0317 09:11:44.435462 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 09:11:44 crc kubenswrapper[4813]: E0317 09:11:44.435467 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 09:11:44 crc kubenswrapper[4813]: E0317 09:11:44.435476 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:44 crc kubenswrapper[4813]: E0317 09:11:44.435476 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 09:11:45.435458942 +0000 UTC m=+127.536262431 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 09:11:44 crc kubenswrapper[4813]: E0317 09:11:44.435571 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 09:11:45.435546245 +0000 UTC m=+127.536349744 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 09:11:44 crc kubenswrapper[4813]: E0317 09:11:44.435632 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 09:11:44 crc kubenswrapper[4813]: E0317 09:11:44.435644 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-17 09:11:45.435580006 +0000 UTC m=+127.536383505 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:44 crc kubenswrapper[4813]: E0317 09:11:44.435685 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs podName:27d4c184-5dd3-492a-b927-f7a7f31291ef nodeName:}" failed. No retries permitted until 2026-03-17 09:11:45.435655179 +0000 UTC m=+127.536458698 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs") pod "network-metrics-daemon-l47ql" (UID: "27d4c184-5dd3-492a-b927-f7a7f31291ef") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 09:11:44 crc kubenswrapper[4813]: E0317 09:11:44.436361 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 09:11:44 crc kubenswrapper[4813]: E0317 09:11:44.436393 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 09:11:44 crc kubenswrapper[4813]: E0317 09:11:44.436406 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:44 crc kubenswrapper[4813]: E0317 09:11:44.436471 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-17 09:11:45.436454985 +0000 UTC m=+127.537258474 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.730440 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:11:44 crc kubenswrapper[4813]: E0317 09:11:44.730676 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.735697 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.736341 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.737311 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.738036 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.738788 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.739428 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.740119 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.742259 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.743036 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.744188 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.744901 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.746203 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.746763 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.747342 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.748337 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.748887 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.749845 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.750237 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.750958 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.752279 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.753137 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.753854 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.754847 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.755695 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.756784 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.757517 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.758699 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.759231 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.760286 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.760878 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.761372 4813 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.761916 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.763666 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.764284 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.765278 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.767189 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.768345 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.769543 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.770464 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.771832 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.772491 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.773790 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.774448 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.775388 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.776179 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.777547 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.778498 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.779892 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.780416 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.781623 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.782177 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.782750 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.784053 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 17 09:11:44 crc kubenswrapper[4813]: I0317 09:11:44.784582 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.267574 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5jjhb" event={"ID":"c1246d4d-93d5-4a97-bef8-1ed881e1a217","Type":"ContainerStarted","Data":"ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737"} Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.267646 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5jjhb" event={"ID":"c1246d4d-93d5-4a97-bef8-1ed881e1a217","Type":"ContainerStarted","Data":"efeb97f54353449254945237e34ed0adebbd28c0ad922eeea06bed59539ee5e4"} Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.270760 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d"} Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.270806 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423"} Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.272667 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-62stj" event={"ID":"8a42f832-caac-4d43-aba1-63eab74eb5f5","Type":"ContainerStarted","Data":"845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db"} Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.273841 4813 generic.go:334] "Generic (PLEG): container finished" podID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerID="06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5" exitCode=0 Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.273906 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerDied","Data":"06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5"} Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.276089 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerStarted","Data":"290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be"} Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.276121 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerStarted","Data":"1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09"} Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.278124 4813 generic.go:334] "Generic (PLEG): container finished" podID="d35d2ade-2829-4fc2-864d-1871c18006f7" containerID="5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d" exitCode=0 Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.278181 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" event={"ID":"d35d2ade-2829-4fc2-864d-1871c18006f7","Type":"ContainerDied","Data":"5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d"} Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.279891 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-zr9rn" event={"ID":"95fc4db5-3177-485a-adfd-f829495e487a","Type":"ContainerStarted","Data":"c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe"} Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.281647 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a"} Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.285570 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" event={"ID":"8ac387df-f9b1-4ce9-a109-1b80e7659f2e","Type":"ContainerStarted","Data":"ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2"} Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.285651 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" event={"ID":"8ac387df-f9b1-4ce9-a109-1b80e7659f2e","Type":"ContainerStarted","Data":"5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68"} Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.288748 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.315095 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.346873 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:11:45 crc kubenswrapper[4813]: E0317 09:11:45.347049 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:11:47.34702272 +0000 UTC m=+129.447826219 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.364869 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.382097 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.401575 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.418838 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.431885 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.444007 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.448134 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.448166 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.448194 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.448213 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.448230 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs\") pod \"network-metrics-daemon-l47ql\" (UID: \"27d4c184-5dd3-492a-b927-f7a7f31291ef\") " pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:11:45 crc kubenswrapper[4813]: E0317 09:11:45.448319 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 09:11:45 crc kubenswrapper[4813]: E0317 09:11:45.448355 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs podName:27d4c184-5dd3-492a-b927-f7a7f31291ef nodeName:}" failed. No retries permitted until 2026-03-17 09:11:47.448344022 +0000 UTC m=+129.549147521 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs") pod "network-metrics-daemon-l47ql" (UID: "27d4c184-5dd3-492a-b927-f7a7f31291ef") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 09:11:45 crc kubenswrapper[4813]: E0317 09:11:45.448569 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 09:11:45 crc kubenswrapper[4813]: E0317 09:11:45.448588 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 09:11:45 crc kubenswrapper[4813]: E0317 09:11:45.448614 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:45 crc kubenswrapper[4813]: E0317 09:11:45.448640 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-17 09:11:47.448632792 +0000 UTC m=+129.549436291 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:45 crc kubenswrapper[4813]: E0317 09:11:45.448677 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 09:11:45 crc kubenswrapper[4813]: E0317 09:11:45.448698 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 09:11:47.448692904 +0000 UTC m=+129.549496403 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 09:11:45 crc kubenswrapper[4813]: E0317 09:11:45.448741 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 09:11:45 crc kubenswrapper[4813]: E0317 09:11:45.448749 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 09:11:45 crc kubenswrapper[4813]: E0317 09:11:45.448755 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:45 crc kubenswrapper[4813]: E0317 09:11:45.448773 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-17 09:11:47.448767706 +0000 UTC m=+129.549571205 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:45 crc kubenswrapper[4813]: E0317 09:11:45.448800 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 09:11:45 crc kubenswrapper[4813]: E0317 09:11:45.448818 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 09:11:47.448812508 +0000 UTC m=+129.549616007 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.457197 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.468387 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.493433 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.505946 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.520146 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.546530 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.559449 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.572830 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.587319 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.602748 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.616086 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.626636 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.642362 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.656091 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.670439 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.685314 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.697455 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.708557 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.720047 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.730394 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:11:45 crc kubenswrapper[4813]: E0317 09:11:45.730510 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.730567 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:11:45 crc kubenswrapper[4813]: E0317 09:11:45.730635 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.730670 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:11:45 crc kubenswrapper[4813]: E0317 09:11:45.730708 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:11:45 crc kubenswrapper[4813]: I0317 09:11:45.735508 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:45Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:46 crc kubenswrapper[4813]: I0317 09:11:46.295506 4813 generic.go:334] "Generic (PLEG): container finished" podID="d35d2ade-2829-4fc2-864d-1871c18006f7" containerID="d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7" exitCode=0 Mar 17 09:11:46 crc kubenswrapper[4813]: I0317 09:11:46.295690 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" event={"ID":"d35d2ade-2829-4fc2-864d-1871c18006f7","Type":"ContainerDied","Data":"d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7"} Mar 17 09:11:46 crc kubenswrapper[4813]: I0317 09:11:46.304586 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerStarted","Data":"dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd"} Mar 17 09:11:46 crc kubenswrapper[4813]: I0317 09:11:46.304675 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerStarted","Data":"0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded"} Mar 17 09:11:46 crc kubenswrapper[4813]: I0317 09:11:46.304690 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerStarted","Data":"791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed"} Mar 17 09:11:46 crc kubenswrapper[4813]: I0317 09:11:46.304702 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerStarted","Data":"7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7"} Mar 17 09:11:46 crc kubenswrapper[4813]: I0317 09:11:46.309736 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:46Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:46 crc kubenswrapper[4813]: I0317 09:11:46.321245 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:46Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:46 crc kubenswrapper[4813]: I0317 09:11:46.333587 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:46Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:46 crc kubenswrapper[4813]: I0317 09:11:46.344621 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:46Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:46 crc kubenswrapper[4813]: I0317 09:11:46.359476 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:46Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:46 crc kubenswrapper[4813]: I0317 09:11:46.378280 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:46Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:46 crc kubenswrapper[4813]: I0317 09:11:46.396549 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:46Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:46 crc kubenswrapper[4813]: I0317 09:11:46.413326 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:46Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:46 crc kubenswrapper[4813]: I0317 09:11:46.429756 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:46Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:46 crc kubenswrapper[4813]: I0317 09:11:46.449294 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:46Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:46 crc kubenswrapper[4813]: I0317 09:11:46.472047 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:46Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:46 crc kubenswrapper[4813]: I0317 09:11:46.483002 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:46Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:46 crc kubenswrapper[4813]: I0317 09:11:46.495359 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:46Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:46 crc kubenswrapper[4813]: I0317 09:11:46.509660 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:46Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:46 crc kubenswrapper[4813]: I0317 09:11:46.729916 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:11:46 crc kubenswrapper[4813]: E0317 09:11:46.730369 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.309520 4813 generic.go:334] "Generic (PLEG): container finished" podID="d35d2ade-2829-4fc2-864d-1871c18006f7" containerID="998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4" exitCode=0 Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.310486 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" event={"ID":"d35d2ade-2829-4fc2-864d-1871c18006f7","Type":"ContainerDied","Data":"998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4"} Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.312350 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9"} Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.317625 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerStarted","Data":"41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b"} Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.317672 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerStarted","Data":"27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923"} Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.330793 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.348385 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.365728 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.371462 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:11:47 crc kubenswrapper[4813]: E0317 09:11:47.371828 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:11:51.371779068 +0000 UTC m=+133.472582607 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.381630 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.399404 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.414064 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.436007 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.456949 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.472752 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.472813 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs\") pod \"network-metrics-daemon-l47ql\" (UID: \"27d4c184-5dd3-492a-b927-f7a7f31291ef\") " pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.472853 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.472917 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.472955 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:11:47 crc kubenswrapper[4813]: E0317 09:11:47.472961 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 09:11:47 crc kubenswrapper[4813]: E0317 09:11:47.472993 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 09:11:47 crc kubenswrapper[4813]: E0317 09:11:47.473007 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:47 crc kubenswrapper[4813]: E0317 09:11:47.473097 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-17 09:11:51.473077468 +0000 UTC m=+133.573881007 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:47 crc kubenswrapper[4813]: E0317 09:11:47.473360 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 09:11:47 crc kubenswrapper[4813]: E0317 09:11:47.473377 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 09:11:47 crc kubenswrapper[4813]: E0317 09:11:47.473385 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:47 crc kubenswrapper[4813]: E0317 09:11:47.473417 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-17 09:11:51.473405679 +0000 UTC m=+133.574209248 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:47 crc kubenswrapper[4813]: E0317 09:11:47.473476 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 09:11:47 crc kubenswrapper[4813]: E0317 09:11:47.473507 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs podName:27d4c184-5dd3-492a-b927-f7a7f31291ef nodeName:}" failed. No retries permitted until 2026-03-17 09:11:51.473498432 +0000 UTC m=+133.574302051 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs") pod "network-metrics-daemon-l47ql" (UID: "27d4c184-5dd3-492a-b927-f7a7f31291ef") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 09:11:47 crc kubenswrapper[4813]: E0317 09:11:47.473549 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 09:11:47 crc kubenswrapper[4813]: E0317 09:11:47.473579 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 09:11:51.473571135 +0000 UTC m=+133.574374744 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 09:11:47 crc kubenswrapper[4813]: E0317 09:11:47.473839 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 09:11:47 crc kubenswrapper[4813]: E0317 09:11:47.473978 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 09:11:51.473950976 +0000 UTC m=+133.574754515 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.474563 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.489820 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.505289 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.523534 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.540110 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.552348 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.572233 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.586427 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.601195 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.615284 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.628326 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.639582 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.655646 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.674200 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.688544 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.707191 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.722047 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.730472 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.730546 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.730474 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:11:47 crc kubenswrapper[4813]: E0317 09:11:47.730665 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:11:47 crc kubenswrapper[4813]: E0317 09:11:47.730696 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:11:47 crc kubenswrapper[4813]: E0317 09:11:47.730793 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.739930 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.755067 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:47 crc kubenswrapper[4813]: I0317 09:11:47.765616 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:47Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.058510 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.058582 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.058629 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.058661 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.058687 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:48Z","lastTransitionTime":"2026-03-17T09:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:48 crc kubenswrapper[4813]: E0317 09:11:48.079440 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.084172 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.084220 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.084237 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.084259 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.084277 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:48Z","lastTransitionTime":"2026-03-17T09:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:48 crc kubenswrapper[4813]: E0317 09:11:48.103778 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.109075 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.109138 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.109156 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.109185 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.109205 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:48Z","lastTransitionTime":"2026-03-17T09:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:48 crc kubenswrapper[4813]: E0317 09:11:48.129638 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.134810 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.134873 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.134893 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.134919 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.134936 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:48Z","lastTransitionTime":"2026-03-17T09:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:48 crc kubenswrapper[4813]: E0317 09:11:48.152502 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.157236 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.157287 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.157307 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.157329 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.157347 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:48Z","lastTransitionTime":"2026-03-17T09:11:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:48 crc kubenswrapper[4813]: E0317 09:11:48.177951 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: E0317 09:11:48.178172 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.324810 4813 generic.go:334] "Generic (PLEG): container finished" podID="d35d2ade-2829-4fc2-864d-1871c18006f7" containerID="d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b" exitCode=0 Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.324895 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" event={"ID":"d35d2ade-2829-4fc2-864d-1871c18006f7","Type":"ContainerDied","Data":"d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b"} Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.350060 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.371709 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.390196 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.408049 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.432438 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.448938 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.468167 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.485124 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.500305 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.517252 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.529394 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.548971 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.562086 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.578559 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.730253 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:11:48 crc kubenswrapper[4813]: E0317 09:11:48.730395 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.745380 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.757857 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.773049 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.783969 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.794819 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.805834 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.816482 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.830037 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.842681 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.867199 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: E0317 09:11:48.867370 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.890636 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.905354 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.932200 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:48 crc kubenswrapper[4813]: I0317 09:11:48.944820 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.335037 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerStarted","Data":"69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f"} Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.339139 4813 generic.go:334] "Generic (PLEG): container finished" podID="d35d2ade-2829-4fc2-864d-1871c18006f7" containerID="f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6" exitCode=0 Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.339361 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" event={"ID":"d35d2ade-2829-4fc2-864d-1871c18006f7","Type":"ContainerDied","Data":"f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6"} Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.361844 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.386134 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.399409 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.415360 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.429322 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.441333 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.458348 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.473512 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.488898 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.509193 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.528157 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.553319 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.565211 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.568935 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.576886 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.577348 4813 scope.go:117] "RemoveContainer" containerID="05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f" Mar 17 09:11:49 crc kubenswrapper[4813]: E0317 09:11:49.577701 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.582175 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.595335 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.608616 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.626547 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.642227 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.660278 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.679973 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.693018 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.706719 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.718860 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.728945 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.730111 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.730150 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.730191 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:11:49 crc kubenswrapper[4813]: E0317 09:11:49.730536 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:11:49 crc kubenswrapper[4813]: E0317 09:11:49.730641 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:11:49 crc kubenswrapper[4813]: E0317 09:11:49.730727 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.736557 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.744717 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.755653 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.768653 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.781046 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:49 crc kubenswrapper[4813]: I0317 09:11:49.796954 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:50 crc kubenswrapper[4813]: I0317 09:11:50.350029 4813 generic.go:334] "Generic (PLEG): container finished" podID="d35d2ade-2829-4fc2-864d-1871c18006f7" containerID="ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f" exitCode=0 Mar 17 09:11:50 crc kubenswrapper[4813]: I0317 09:11:50.350139 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" event={"ID":"d35d2ade-2829-4fc2-864d-1871c18006f7","Type":"ContainerDied","Data":"ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f"} Mar 17 09:11:50 crc kubenswrapper[4813]: I0317 09:11:50.351507 4813 scope.go:117] "RemoveContainer" containerID="05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f" Mar 17 09:11:50 crc kubenswrapper[4813]: E0317 09:11:50.351824 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:11:50 crc kubenswrapper[4813]: I0317 09:11:50.373264 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:50Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:50 crc kubenswrapper[4813]: I0317 09:11:50.392675 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:50Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:50 crc kubenswrapper[4813]: I0317 09:11:50.416691 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:50Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:50 crc kubenswrapper[4813]: I0317 09:11:50.436247 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:50Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:50 crc kubenswrapper[4813]: I0317 09:11:50.459580 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:50Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:50 crc kubenswrapper[4813]: I0317 09:11:50.474145 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:50Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:50 crc kubenswrapper[4813]: I0317 09:11:50.492304 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:50Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:50 crc kubenswrapper[4813]: I0317 09:11:50.515672 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:50Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:50 crc kubenswrapper[4813]: I0317 09:11:50.542887 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:50Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:50 crc kubenswrapper[4813]: I0317 09:11:50.559211 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:50Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:50 crc kubenswrapper[4813]: I0317 09:11:50.576359 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:50Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:50 crc kubenswrapper[4813]: I0317 09:11:50.592455 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:50Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:50 crc kubenswrapper[4813]: I0317 09:11:50.606243 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:50Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:50 crc kubenswrapper[4813]: I0317 09:11:50.617672 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:50Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:50 crc kubenswrapper[4813]: I0317 09:11:50.629045 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:50Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:50 crc kubenswrapper[4813]: I0317 09:11:50.641903 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:50Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:50 crc kubenswrapper[4813]: I0317 09:11:50.730261 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:11:50 crc kubenswrapper[4813]: E0317 09:11:50.730435 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.359677 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerStarted","Data":"3bffe0e09f7091172999ee64de77b3f191727eeeb12e27962b669504d43d6adb"} Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.360192 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.360230 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.360257 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.366246 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" event={"ID":"d35d2ade-2829-4fc2-864d-1871c18006f7","Type":"ContainerStarted","Data":"238868fbd0e0229eec91acd7a54dcd22c659471003620351cba19de858aef8ac"} Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.384881 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.406404 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.408508 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.411567 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.415389 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:11:51 crc kubenswrapper[4813]: E0317 09:11:51.416249 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:11:59.416140704 +0000 UTC m=+141.516944253 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.427077 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.451853 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.471361 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.507149 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bffe0e09f7091172999ee64de77b3f191727eeeb12e27962b669504d43d6adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.517032 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.517086 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.517177 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.517238 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:11:51 crc kubenswrapper[4813]: E0317 09:11:51.517261 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.517272 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs\") pod \"network-metrics-daemon-l47ql\" (UID: \"27d4c184-5dd3-492a-b927-f7a7f31291ef\") " pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:11:51 crc kubenswrapper[4813]: E0317 09:11:51.517385 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 09:11:51 crc kubenswrapper[4813]: E0317 09:11:51.517427 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 09:11:51 crc kubenswrapper[4813]: E0317 09:11:51.517446 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:51 crc kubenswrapper[4813]: E0317 09:11:51.517440 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 09:11:51 crc kubenswrapper[4813]: E0317 09:11:51.517403 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 09:11:59.517362241 +0000 UTC m=+141.618165780 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 09:11:51 crc kubenswrapper[4813]: E0317 09:11:51.517395 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 09:11:51 crc kubenswrapper[4813]: E0317 09:11:51.517579 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs podName:27d4c184-5dd3-492a-b927-f7a7f31291ef nodeName:}" failed. No retries permitted until 2026-03-17 09:11:59.517557017 +0000 UTC m=+141.618360546 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs") pod "network-metrics-daemon-l47ql" (UID: "27d4c184-5dd3-492a-b927-f7a7f31291ef") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 09:11:51 crc kubenswrapper[4813]: E0317 09:11:51.517795 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 09:11:59.517754534 +0000 UTC m=+141.618558063 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 09:11:51 crc kubenswrapper[4813]: E0317 09:11:51.517831 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-17 09:11:59.517814856 +0000 UTC m=+141.618618485 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:51 crc kubenswrapper[4813]: E0317 09:11:51.518081 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 09:11:51 crc kubenswrapper[4813]: E0317 09:11:51.518182 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 09:11:51 crc kubenswrapper[4813]: E0317 09:11:51.518288 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:51 crc kubenswrapper[4813]: E0317 09:11:51.518437 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-17 09:11:59.518423887 +0000 UTC m=+141.619227406 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.523811 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.542004 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.557375 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.574790 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.594402 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.604489 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.612846 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.621120 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.632475 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.640884 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.650418 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.661450 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.670526 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.688111 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.702447 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.711729 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.729850 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:11:51 crc kubenswrapper[4813]: E0317 09:11:51.729969 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.730257 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:11:51 crc kubenswrapper[4813]: E0317 09:11:51.730307 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.730342 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:11:51 crc kubenswrapper[4813]: E0317 09:11:51.730383 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.730780 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238868fbd0e0229eec91acd7a54dcd22c659471003620351cba19de858aef8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.742845 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.768405 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bffe0e09f7091172999ee64de77b3f191727eeeb12e27962b669504d43d6adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.782579 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.802055 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.822099 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.844338 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.858911 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.879122 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:51 crc kubenswrapper[4813]: I0317 09:11:51.895347 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:51Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:52 crc kubenswrapper[4813]: I0317 09:11:52.083884 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:11:52 crc kubenswrapper[4813]: I0317 09:11:52.084878 4813 scope.go:117] "RemoveContainer" containerID="05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f" Mar 17 09:11:52 crc kubenswrapper[4813]: E0317 09:11:52.085147 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:11:52 crc kubenswrapper[4813]: I0317 09:11:52.730254 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:11:52 crc kubenswrapper[4813]: E0317 09:11:52.730365 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:11:53 crc kubenswrapper[4813]: I0317 09:11:53.730311 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:11:53 crc kubenswrapper[4813]: I0317 09:11:53.730439 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:11:53 crc kubenswrapper[4813]: E0317 09:11:53.730523 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:11:53 crc kubenswrapper[4813]: I0317 09:11:53.730399 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:11:53 crc kubenswrapper[4813]: E0317 09:11:53.730688 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:11:53 crc kubenswrapper[4813]: E0317 09:11:53.730946 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:11:53 crc kubenswrapper[4813]: E0317 09:11:53.869052 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:11:54 crc kubenswrapper[4813]: I0317 09:11:54.380385 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lngs6_a1730ca2-a1bb-4e9a-ba56-7edaded79f36/ovnkube-controller/0.log" Mar 17 09:11:54 crc kubenswrapper[4813]: I0317 09:11:54.384591 4813 generic.go:334] "Generic (PLEG): container finished" podID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerID="3bffe0e09f7091172999ee64de77b3f191727eeeb12e27962b669504d43d6adb" exitCode=1 Mar 17 09:11:54 crc kubenswrapper[4813]: I0317 09:11:54.384740 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerDied","Data":"3bffe0e09f7091172999ee64de77b3f191727eeeb12e27962b669504d43d6adb"} Mar 17 09:11:54 crc kubenswrapper[4813]: I0317 09:11:54.386649 4813 scope.go:117] "RemoveContainer" containerID="3bffe0e09f7091172999ee64de77b3f191727eeeb12e27962b669504d43d6adb" Mar 17 09:11:54 crc kubenswrapper[4813]: I0317 09:11:54.405769 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:54Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:54 crc kubenswrapper[4813]: I0317 09:11:54.426813 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:54Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:54 crc kubenswrapper[4813]: I0317 09:11:54.440679 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:54Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:54 crc kubenswrapper[4813]: I0317 09:11:54.454155 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:54Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:54 crc kubenswrapper[4813]: I0317 09:11:54.467308 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:54Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:54 crc kubenswrapper[4813]: I0317 09:11:54.479263 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:54Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:54 crc kubenswrapper[4813]: I0317 09:11:54.492426 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:54Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:54 crc kubenswrapper[4813]: I0317 09:11:54.506238 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:54Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:54 crc kubenswrapper[4813]: I0317 09:11:54.518705 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:54Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:54 crc kubenswrapper[4813]: I0317 09:11:54.534228 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238868fbd0e0229eec91acd7a54dcd22c659471003620351cba19de858aef8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:54Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:54 crc kubenswrapper[4813]: I0317 09:11:54.547384 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:54Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:54 crc kubenswrapper[4813]: I0317 09:11:54.560998 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:54Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:54 crc kubenswrapper[4813]: I0317 09:11:54.576433 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:54Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:54 crc kubenswrapper[4813]: I0317 09:11:54.591281 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:54Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:54 crc kubenswrapper[4813]: I0317 09:11:54.607966 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:54Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:54 crc kubenswrapper[4813]: I0317 09:11:54.628175 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3bffe0e09f7091172999ee64de77b3f191727eeeb12e27962b669504d43d6adb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bffe0e09f7091172999ee64de77b3f191727eeeb12e27962b669504d43d6adb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:11:53Z\\\",\\\"message\\\":\\\"ler/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0317 09:11:53.795035 6794 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0317 09:11:53.795088 6794 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0317 09:11:53.795518 6794 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 09:11:53.795713 6794 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 09:11:53.796449 6794 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 09:11:53.796495 6794 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 09:11:53.796503 6794 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 09:11:53.796552 6794 factory.go:656] Stopping watch factory\\\\nI0317 09:11:53.796576 6794 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 09:11:53.796590 6794 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 09:11:53.796630 6794 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:54Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:54 crc kubenswrapper[4813]: I0317 09:11:54.730038 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:11:54 crc kubenswrapper[4813]: E0317 09:11:54.730184 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.390980 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lngs6_a1730ca2-a1bb-4e9a-ba56-7edaded79f36/ovnkube-controller/0.log" Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.394635 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerStarted","Data":"3af8690257686e92e135e44fbad156e0d724f31b8ae62ad636a0954ec8f0f1c6"} Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.395214 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.418590 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:55Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.439341 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:55Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.460014 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:55Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.477753 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:55Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.498092 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:55Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.517215 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:55Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.533268 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:55Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.553261 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:55Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.572539 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:55Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.601198 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238868fbd0e0229eec91acd7a54dcd22c659471003620351cba19de858aef8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:55Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.614674 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:55Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.631787 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:55Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.647049 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:55Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.667913 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:55Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.692051 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:55Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.723956 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8690257686e92e135e44fbad156e0d724f31b8ae62ad636a0954ec8f0f1c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bffe0e09f7091172999ee64de77b3f191727eeeb12e27962b669504d43d6adb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:11:53Z\\\",\\\"message\\\":\\\"ler/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0317 09:11:53.795035 6794 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0317 09:11:53.795088 6794 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0317 09:11:53.795518 6794 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 09:11:53.795713 6794 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 09:11:53.796449 6794 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 09:11:53.796495 6794 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 09:11:53.796503 6794 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 09:11:53.796552 6794 factory.go:656] Stopping watch factory\\\\nI0317 09:11:53.796576 6794 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 09:11:53.796590 6794 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 09:11:53.796630 6794 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:55Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.730005 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.730058 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:11:55 crc kubenswrapper[4813]: I0317 09:11:55.730000 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:11:55 crc kubenswrapper[4813]: E0317 09:11:55.730212 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:11:55 crc kubenswrapper[4813]: E0317 09:11:55.730364 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:11:55 crc kubenswrapper[4813]: E0317 09:11:55.730520 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.401479 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lngs6_a1730ca2-a1bb-4e9a-ba56-7edaded79f36/ovnkube-controller/1.log" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.403850 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lngs6_a1730ca2-a1bb-4e9a-ba56-7edaded79f36/ovnkube-controller/0.log" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.408157 4813 generic.go:334] "Generic (PLEG): container finished" podID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerID="3af8690257686e92e135e44fbad156e0d724f31b8ae62ad636a0954ec8f0f1c6" exitCode=1 Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.408235 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerDied","Data":"3af8690257686e92e135e44fbad156e0d724f31b8ae62ad636a0954ec8f0f1c6"} Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.408307 4813 scope.go:117] "RemoveContainer" containerID="3bffe0e09f7091172999ee64de77b3f191727eeeb12e27962b669504d43d6adb" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.409431 4813 scope.go:117] "RemoveContainer" containerID="3af8690257686e92e135e44fbad156e0d724f31b8ae62ad636a0954ec8f0f1c6" Mar 17 09:11:56 crc kubenswrapper[4813]: E0317 09:11:56.409781 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lngs6_openshift-ovn-kubernetes(a1730ca2-a1bb-4e9a-ba56-7edaded79f36)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.435081 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:56Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.452833 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:56Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.477550 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238868fbd0e0229eec91acd7a54dcd22c659471003620351cba19de858aef8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:56Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.496265 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:56Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.516481 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:56Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.535427 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:56Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.553392 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:56Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.573825 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:56Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.603753 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8690257686e92e135e44fbad156e0d724f31b8ae62ad636a0954ec8f0f1c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3bffe0e09f7091172999ee64de77b3f191727eeeb12e27962b669504d43d6adb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:11:53Z\\\",\\\"message\\\":\\\"ler/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0317 09:11:53.795035 6794 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI0317 09:11:53.795088 6794 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI0317 09:11:53.795518 6794 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 09:11:53.795713 6794 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 09:11:53.796449 6794 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 09:11:53.796495 6794 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 09:11:53.796503 6794 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 09:11:53.796552 6794 factory.go:656] Stopping watch factory\\\\nI0317 09:11:53.796576 6794 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 09:11:53.796590 6794 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 09:11:53.796630 6794 handler.go:208] Removed *v1.Namespace ev\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8690257686e92e135e44fbad156e0d724f31b8ae62ad636a0954ec8f0f1c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:11:55Z\\\",\\\"message\\\":\\\"Set:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0317 09:11:55.334321 6962 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0317 09:11:55.335306 6962 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0317 09:11:55.335419 6962 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:54Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:56Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.625866 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:56Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.643306 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:56Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.661449 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:56Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.675696 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:56Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.688304 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:56Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.702638 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:56Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.719423 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:56Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.734978 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:11:56 crc kubenswrapper[4813]: E0317 09:11:56.735445 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:11:56 crc kubenswrapper[4813]: I0317 09:11:56.753584 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.412227 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lngs6_a1730ca2-a1bb-4e9a-ba56-7edaded79f36/ovnkube-controller/1.log" Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.415338 4813 scope.go:117] "RemoveContainer" containerID="3af8690257686e92e135e44fbad156e0d724f31b8ae62ad636a0954ec8f0f1c6" Mar 17 09:11:57 crc kubenswrapper[4813]: E0317 09:11:57.415478 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lngs6_openshift-ovn-kubernetes(a1730ca2-a1bb-4e9a-ba56-7edaded79f36)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.428536 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:57Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.449201 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:57Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.469576 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:57Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.489863 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:57Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.508529 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:57Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.523968 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:57Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.540493 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:57Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.558865 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:57Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.578731 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:57Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.597981 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:57Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.612949 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:57Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.634194 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238868fbd0e0229eec91acd7a54dcd22c659471003620351cba19de858aef8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:57Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.651324 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:57Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.680383 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8690257686e92e135e44fbad156e0d724f31b8ae62ad636a0954ec8f0f1c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8690257686e92e135e44fbad156e0d724f31b8ae62ad636a0954ec8f0f1c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:11:55Z\\\",\\\"message\\\":\\\"Set:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0317 09:11:55.334321 6962 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0317 09:11:55.335306 6962 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0317 09:11:55.335419 6962 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lngs6_openshift-ovn-kubernetes(a1730ca2-a1bb-4e9a-ba56-7edaded79f36)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:57Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.701592 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ecb2c81-a49d-4d2e-83b2-b821ea50c5e5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80403c9745d70069a406ca4771940c9ff6958417787d62595bf005db27d2c1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a5b5d0590fefa483bcb5d5a78dfee7d61881fe9e19e5f229d3788d2cc09ab87\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:10:07Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0317 09:09:41.118018 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0317 09:09:41.120994 1 observer_polling.go:159] Starting file observer\\\\nI0317 09:09:41.159745 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0317 09:09:41.162501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0317 09:10:07.008700 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0317 09:10:07.008843 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:06Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1d7eaebd9e22a3f68b00c57f0bc9971826fc05a41062947a1ce04f1e8d380c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa97f1c57fb3a119266cee961a8872c9115b7541542c57598511fd47ea3c907\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3147869c6b9a5426228139c412d8b7b2cb224a771ca092fbe9fe3b9b2b215d61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:57Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.716648 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:57Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.730562 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.730580 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:11:57 crc kubenswrapper[4813]: E0317 09:11:57.730757 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.730735 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:11:57 crc kubenswrapper[4813]: E0317 09:11:57.730957 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:11:57 crc kubenswrapper[4813]: E0317 09:11:57.731098 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:11:57 crc kubenswrapper[4813]: I0317 09:11:57.732748 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:57Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.269064 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.269126 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.269159 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.269206 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.269229 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:58Z","lastTransitionTime":"2026-03-17T09:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:58 crc kubenswrapper[4813]: E0317 09:11:58.289678 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.295095 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.295201 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.295225 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.295255 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.295278 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:58Z","lastTransitionTime":"2026-03-17T09:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:58 crc kubenswrapper[4813]: E0317 09:11:58.316011 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.321058 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.321114 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.321135 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.321159 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.321175 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:58Z","lastTransitionTime":"2026-03-17T09:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:58 crc kubenswrapper[4813]: E0317 09:11:58.341055 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.346020 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.346151 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.346179 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.346208 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.346229 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:58Z","lastTransitionTime":"2026-03-17T09:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:58 crc kubenswrapper[4813]: E0317 09:11:58.367808 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.372557 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.372643 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.372662 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.372688 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.372706 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:11:58Z","lastTransitionTime":"2026-03-17T09:11:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:11:58 crc kubenswrapper[4813]: E0317 09:11:58.392095 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:58 crc kubenswrapper[4813]: E0317 09:11:58.392330 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.730074 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:11:58 crc kubenswrapper[4813]: E0317 09:11:58.730880 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.748446 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.768844 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.788124 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.802989 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.820108 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.834997 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.852514 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:58 crc kubenswrapper[4813]: E0317 09:11:58.869521 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.874417 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.891157 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.913311 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238868fbd0e0229eec91acd7a54dcd22c659471003620351cba19de858aef8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.930690 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.949129 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.964161 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ecb2c81-a49d-4d2e-83b2-b821ea50c5e5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80403c9745d70069a406ca4771940c9ff6958417787d62595bf005db27d2c1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a5b5d0590fefa483bcb5d5a78dfee7d61881fe9e19e5f229d3788d2cc09ab87\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:10:07Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0317 09:09:41.118018 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0317 09:09:41.120994 1 observer_polling.go:159] Starting file observer\\\\nI0317 09:09:41.159745 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0317 09:09:41.162501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0317 09:10:07.008700 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0317 09:10:07.008843 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:06Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1d7eaebd9e22a3f68b00c57f0bc9971826fc05a41062947a1ce04f1e8d380c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa97f1c57fb3a119266cee961a8872c9115b7541542c57598511fd47ea3c907\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3147869c6b9a5426228139c412d8b7b2cb224a771ca092fbe9fe3b9b2b215d61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.980024 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:58 crc kubenswrapper[4813]: I0317 09:11:58.994212 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:59 crc kubenswrapper[4813]: I0317 09:11:59.013855 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:59Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:59 crc kubenswrapper[4813]: I0317 09:11:59.046730 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8690257686e92e135e44fbad156e0d724f31b8ae62ad636a0954ec8f0f1c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8690257686e92e135e44fbad156e0d724f31b8ae62ad636a0954ec8f0f1c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:11:55Z\\\",\\\"message\\\":\\\"Set:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0317 09:11:55.334321 6962 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0317 09:11:55.335306 6962 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0317 09:11:55.335419 6962 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lngs6_openshift-ovn-kubernetes(a1730ca2-a1bb-4e9a-ba56-7edaded79f36)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:11:59Z is after 2025-08-24T17:21:41Z" Mar 17 09:11:59 crc kubenswrapper[4813]: I0317 09:11:59.465720 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:11:59 crc kubenswrapper[4813]: E0317 09:11:59.466039 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:12:15.466012702 +0000 UTC m=+157.566816241 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:11:59 crc kubenswrapper[4813]: I0317 09:11:59.566924 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:11:59 crc kubenswrapper[4813]: I0317 09:11:59.566989 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:11:59 crc kubenswrapper[4813]: I0317 09:11:59.567024 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs\") pod \"network-metrics-daemon-l47ql\" (UID: \"27d4c184-5dd3-492a-b927-f7a7f31291ef\") " pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:11:59 crc kubenswrapper[4813]: I0317 09:11:59.567075 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:11:59 crc kubenswrapper[4813]: I0317 09:11:59.567111 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:11:59 crc kubenswrapper[4813]: E0317 09:11:59.567260 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 09:11:59 crc kubenswrapper[4813]: E0317 09:11:59.567311 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 09:11:59 crc kubenswrapper[4813]: E0317 09:11:59.567329 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 09:11:59 crc kubenswrapper[4813]: E0317 09:11:59.567360 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 09:11:59 crc kubenswrapper[4813]: E0317 09:11:59.567542 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 09:11:59 crc kubenswrapper[4813]: E0317 09:11:59.567401 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 09:12:15.567322223 +0000 UTC m=+157.668125762 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 09:11:59 crc kubenswrapper[4813]: E0317 09:11:59.567566 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 09:11:59 crc kubenswrapper[4813]: E0317 09:11:59.567586 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:59 crc kubenswrapper[4813]: E0317 09:11:59.567408 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 09:11:59 crc kubenswrapper[4813]: E0317 09:11:59.567672 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 09:12:15.567574331 +0000 UTC m=+157.668377870 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 09:11:59 crc kubenswrapper[4813]: E0317 09:11:59.567703 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs podName:27d4c184-5dd3-492a-b927-f7a7f31291ef nodeName:}" failed. No retries permitted until 2026-03-17 09:12:15.567688815 +0000 UTC m=+157.668492344 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs") pod "network-metrics-daemon-l47ql" (UID: "27d4c184-5dd3-492a-b927-f7a7f31291ef") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 09:11:59 crc kubenswrapper[4813]: E0317 09:11:59.567791 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-17 09:12:15.567776648 +0000 UTC m=+157.668580187 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:59 crc kubenswrapper[4813]: E0317 09:11:59.567717 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:59 crc kubenswrapper[4813]: E0317 09:11:59.567959 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-17 09:12:15.567897011 +0000 UTC m=+157.668700550 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:11:59 crc kubenswrapper[4813]: I0317 09:11:59.729887 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:11:59 crc kubenswrapper[4813]: I0317 09:11:59.729890 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:11:59 crc kubenswrapper[4813]: I0317 09:11:59.729963 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:11:59 crc kubenswrapper[4813]: E0317 09:11:59.730694 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:11:59 crc kubenswrapper[4813]: E0317 09:11:59.730590 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:11:59 crc kubenswrapper[4813]: E0317 09:11:59.731373 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:00 crc kubenswrapper[4813]: I0317 09:12:00.730500 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:00 crc kubenswrapper[4813]: E0317 09:12:00.730750 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:01 crc kubenswrapper[4813]: I0317 09:12:01.730006 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:01 crc kubenswrapper[4813]: I0317 09:12:01.730080 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:01 crc kubenswrapper[4813]: I0317 09:12:01.730117 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:01 crc kubenswrapper[4813]: E0317 09:12:01.730193 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:01 crc kubenswrapper[4813]: E0317 09:12:01.730403 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:01 crc kubenswrapper[4813]: E0317 09:12:01.730530 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:02 crc kubenswrapper[4813]: I0317 09:12:02.729872 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:02 crc kubenswrapper[4813]: E0317 09:12:02.730074 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:03 crc kubenswrapper[4813]: I0317 09:12:03.731091 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:03 crc kubenswrapper[4813]: I0317 09:12:03.731216 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:03 crc kubenswrapper[4813]: E0317 09:12:03.731297 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:03 crc kubenswrapper[4813]: I0317 09:12:03.731087 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:03 crc kubenswrapper[4813]: E0317 09:12:03.731517 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:03 crc kubenswrapper[4813]: E0317 09:12:03.731688 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:03 crc kubenswrapper[4813]: E0317 09:12:03.871419 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:12:04 crc kubenswrapper[4813]: I0317 09:12:04.730293 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:04 crc kubenswrapper[4813]: E0317 09:12:04.730489 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:05 crc kubenswrapper[4813]: I0317 09:12:05.729898 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:05 crc kubenswrapper[4813]: I0317 09:12:05.729950 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:05 crc kubenswrapper[4813]: I0317 09:12:05.730114 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:05 crc kubenswrapper[4813]: E0317 09:12:05.730305 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:05 crc kubenswrapper[4813]: E0317 09:12:05.730440 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:05 crc kubenswrapper[4813]: E0317 09:12:05.730872 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:06 crc kubenswrapper[4813]: I0317 09:12:06.729963 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:06 crc kubenswrapper[4813]: E0317 09:12:06.730460 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:06 crc kubenswrapper[4813]: I0317 09:12:06.730880 4813 scope.go:117] "RemoveContainer" containerID="05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f" Mar 17 09:12:06 crc kubenswrapper[4813]: E0317 09:12:06.731351 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:12:07 crc kubenswrapper[4813]: I0317 09:12:07.730518 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:07 crc kubenswrapper[4813]: I0317 09:12:07.730651 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:07 crc kubenswrapper[4813]: E0317 09:12:07.730747 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:07 crc kubenswrapper[4813]: I0317 09:12:07.730664 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:07 crc kubenswrapper[4813]: E0317 09:12:07.730886 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:07 crc kubenswrapper[4813]: E0317 09:12:07.730989 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.579227 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.579290 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.579312 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.579342 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.579363 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:08Z","lastTransitionTime":"2026-03-17T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:08 crc kubenswrapper[4813]: E0317 09:12:08.600908 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.606960 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.607015 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.607033 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.607059 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.607076 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:08Z","lastTransitionTime":"2026-03-17T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:08 crc kubenswrapper[4813]: E0317 09:12:08.625754 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.630994 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.631080 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.631104 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.631136 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.631160 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:08Z","lastTransitionTime":"2026-03-17T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:08 crc kubenswrapper[4813]: E0317 09:12:08.649844 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.654835 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.654883 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.654900 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.654924 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.654942 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:08Z","lastTransitionTime":"2026-03-17T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:08 crc kubenswrapper[4813]: E0317 09:12:08.675392 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.681547 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.681625 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.681641 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.681662 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.681676 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:08Z","lastTransitionTime":"2026-03-17T09:12:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:08 crc kubenswrapper[4813]: E0317 09:12:08.701086 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:08 crc kubenswrapper[4813]: E0317 09:12:08.701418 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.730161 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:08 crc kubenswrapper[4813]: E0317 09:12:08.730404 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.752470 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238868fbd0e0229eec91acd7a54dcd22c659471003620351cba19de858aef8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.770293 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.788191 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.807312 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.823262 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.839182 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.856802 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:08 crc kubenswrapper[4813]: E0317 09:12:08.872908 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.894290 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3af8690257686e92e135e44fbad156e0d724f31b8ae62ad636a0954ec8f0f1c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8690257686e92e135e44fbad156e0d724f31b8ae62ad636a0954ec8f0f1c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:11:55Z\\\",\\\"message\\\":\\\"Set:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0317 09:11:55.334321 6962 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0317 09:11:55.335306 6962 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0317 09:11:55.335419 6962 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lngs6_openshift-ovn-kubernetes(a1730ca2-a1bb-4e9a-ba56-7edaded79f36)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.912766 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ecb2c81-a49d-4d2e-83b2-b821ea50c5e5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80403c9745d70069a406ca4771940c9ff6958417787d62595bf005db27d2c1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a5b5d0590fefa483bcb5d5a78dfee7d61881fe9e19e5f229d3788d2cc09ab87\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:10:07Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0317 09:09:41.118018 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0317 09:09:41.120994 1 observer_polling.go:159] Starting file observer\\\\nI0317 09:09:41.159745 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0317 09:09:41.162501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0317 09:10:07.008700 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0317 09:10:07.008843 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:06Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1d7eaebd9e22a3f68b00c57f0bc9971826fc05a41062947a1ce04f1e8d380c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa97f1c57fb3a119266cee961a8872c9115b7541542c57598511fd47ea3c907\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3147869c6b9a5426228139c412d8b7b2cb224a771ca092fbe9fe3b9b2b215d61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.925432 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.940474 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.951219 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.964639 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.977232 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.988975 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:08 crc kubenswrapper[4813]: I0317 09:12:08.999504 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:08Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:09 crc kubenswrapper[4813]: I0317 09:12:09.009224 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:09Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:09 crc kubenswrapper[4813]: I0317 09:12:09.729920 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:09 crc kubenswrapper[4813]: I0317 09:12:09.730046 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:09 crc kubenswrapper[4813]: E0317 09:12:09.730080 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:09 crc kubenswrapper[4813]: I0317 09:12:09.729933 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:09 crc kubenswrapper[4813]: E0317 09:12:09.730241 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:09 crc kubenswrapper[4813]: E0317 09:12:09.730432 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:10 crc kubenswrapper[4813]: I0317 09:12:10.730031 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:10 crc kubenswrapper[4813]: E0317 09:12:10.730278 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:11 crc kubenswrapper[4813]: I0317 09:12:11.729674 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:11 crc kubenswrapper[4813]: I0317 09:12:11.729752 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:11 crc kubenswrapper[4813]: I0317 09:12:11.729711 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:11 crc kubenswrapper[4813]: E0317 09:12:11.729902 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:11 crc kubenswrapper[4813]: E0317 09:12:11.730061 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:11 crc kubenswrapper[4813]: E0317 09:12:11.730261 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:12 crc kubenswrapper[4813]: I0317 09:12:12.729836 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:12 crc kubenswrapper[4813]: E0317 09:12:12.730058 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:12 crc kubenswrapper[4813]: I0317 09:12:12.731269 4813 scope.go:117] "RemoveContainer" containerID="3af8690257686e92e135e44fbad156e0d724f31b8ae62ad636a0954ec8f0f1c6" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.510441 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lngs6_a1730ca2-a1bb-4e9a-ba56-7edaded79f36/ovnkube-controller/1.log" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.514975 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerStarted","Data":"1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f"} Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.515464 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.538942 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:13Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.554227 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:13Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.581412 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:13Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.603575 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:13Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.616681 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:13Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.630184 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:13Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.644088 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:13Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.664674 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238868fbd0e0229eec91acd7a54dcd22c659471003620351cba19de858aef8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:13Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.675818 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:13Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.686559 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:13Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.702139 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:13Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.714053 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:13Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.725231 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:13Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.730155 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.730288 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:13 crc kubenswrapper[4813]: E0317 09:12:13.730477 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.730525 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:13 crc kubenswrapper[4813]: E0317 09:12:13.730756 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:13 crc kubenswrapper[4813]: E0317 09:12:13.730927 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.738486 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:13Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.764183 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8690257686e92e135e44fbad156e0d724f31b8ae62ad636a0954ec8f0f1c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:11:55Z\\\",\\\"message\\\":\\\"Set:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0317 09:11:55.334321 6962 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0317 09:11:55.335306 6962 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0317 09:11:55.335419 6962 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:13Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.776652 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ecb2c81-a49d-4d2e-83b2-b821ea50c5e5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80403c9745d70069a406ca4771940c9ff6958417787d62595bf005db27d2c1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a5b5d0590fefa483bcb5d5a78dfee7d61881fe9e19e5f229d3788d2cc09ab87\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:10:07Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0317 09:09:41.118018 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0317 09:09:41.120994 1 observer_polling.go:159] Starting file observer\\\\nI0317 09:09:41.159745 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0317 09:09:41.162501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0317 09:10:07.008700 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0317 09:10:07.008843 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:06Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1d7eaebd9e22a3f68b00c57f0bc9971826fc05a41062947a1ce04f1e8d380c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa97f1c57fb3a119266cee961a8872c9115b7541542c57598511fd47ea3c907\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3147869c6b9a5426228139c412d8b7b2cb224a771ca092fbe9fe3b9b2b215d61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:13Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:13 crc kubenswrapper[4813]: I0317 09:12:13.792961 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:13Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:13 crc kubenswrapper[4813]: E0317 09:12:13.874643 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.520538 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lngs6_a1730ca2-a1bb-4e9a-ba56-7edaded79f36/ovnkube-controller/2.log" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.521490 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lngs6_a1730ca2-a1bb-4e9a-ba56-7edaded79f36/ovnkube-controller/1.log" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.526232 4813 generic.go:334] "Generic (PLEG): container finished" podID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerID="1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f" exitCode=1 Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.526287 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerDied","Data":"1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f"} Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.526340 4813 scope.go:117] "RemoveContainer" containerID="3af8690257686e92e135e44fbad156e0d724f31b8ae62ad636a0954ec8f0f1c6" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.528506 4813 scope.go:117] "RemoveContainer" containerID="1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f" Mar 17 09:12:14 crc kubenswrapper[4813]: E0317 09:12:14.530823 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lngs6_openshift-ovn-kubernetes(a1730ca2-a1bb-4e9a-ba56-7edaded79f36)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.543746 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:14Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.557989 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:14Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.571675 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:14Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.600124 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3af8690257686e92e135e44fbad156e0d724f31b8ae62ad636a0954ec8f0f1c6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:11:55Z\\\",\\\"message\\\":\\\"Set:[{GoUUID:61897e97-c771-4738-8709-09636387cb00}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7e8bb06a-06a5-45bc-a752-26a17d322811}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0317 09:11:55.334321 6962 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0317 09:11:55.335306 6962 model_client.go:382] Update operations generated as: [{Op:update Table:Logical_Switch_Port Row:map[addresses:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]} options:{GoMap:map[iface-id-ver:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 requested-chassis:crc]} port_security:{GoSet:[0a:58:0a:d9:00:5c 10.217.0.92]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c94130be-172c-477c-88c4-40cc7eba30fe}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF0317 09:11:55.335419 6962 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:54Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:13Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 09:12:13.643423 7150 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0317 09:12:13.643461 7150 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0317 09:12:13.643501 7150 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 09:12:13.643512 7150 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 09:12:13.643511 7150 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 09:12:13.643528 7150 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 09:12:13.643536 7150 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 09:12:13.643582 7150 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 09:12:13.643623 7150 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 09:12:13.643624 7150 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0317 09:12:13.643639 7150 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 09:12:13.643659 7150 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0317 09:12:13.643684 7150 factory.go:656] Stopping watch factory\\\\nI0317 09:12:13.643691 7150 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0317 09:12:13.643701 7150 ovnkube.go:599] Stopped ovnkube\\\\nI0317 09:12:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:12:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:14Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.613780 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ecb2c81-a49d-4d2e-83b2-b821ea50c5e5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80403c9745d70069a406ca4771940c9ff6958417787d62595bf005db27d2c1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a5b5d0590fefa483bcb5d5a78dfee7d61881fe9e19e5f229d3788d2cc09ab87\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:10:07Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0317 09:09:41.118018 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0317 09:09:41.120994 1 observer_polling.go:159] Starting file observer\\\\nI0317 09:09:41.159745 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0317 09:09:41.162501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0317 09:10:07.008700 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0317 09:10:07.008843 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:06Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1d7eaebd9e22a3f68b00c57f0bc9971826fc05a41062947a1ce04f1e8d380c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa97f1c57fb3a119266cee961a8872c9115b7541542c57598511fd47ea3c907\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3147869c6b9a5426228139c412d8b7b2cb224a771ca092fbe9fe3b9b2b215d61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:14Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.625529 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:14Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.643232 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:14Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.653401 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:14Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.666956 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:14Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.679187 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:14Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.689896 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:14Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.700100 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:14Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.718683 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:14Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.730453 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:14 crc kubenswrapper[4813]: E0317 09:12:14.730573 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.738811 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238868fbd0e0229eec91acd7a54dcd22c659471003620351cba19de858aef8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:14Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.759556 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:14Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.779008 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:14Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:14 crc kubenswrapper[4813]: I0317 09:12:14.798727 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:14Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.534031 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lngs6_a1730ca2-a1bb-4e9a-ba56-7edaded79f36/ovnkube-controller/2.log" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.539735 4813 scope.go:117] "RemoveContainer" containerID="1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f" Mar 17 09:12:15 crc kubenswrapper[4813]: E0317 09:12:15.540094 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lngs6_openshift-ovn-kubernetes(a1730ca2-a1bb-4e9a-ba56-7edaded79f36)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.549324 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:12:15 crc kubenswrapper[4813]: E0317 09:12:15.549486 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:12:47.549452594 +0000 UTC m=+189.650256123 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.566271 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:15Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.588045 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:15Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.606995 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:15Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.620588 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:15Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.638459 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:15Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.650250 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:15Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.650658 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.650720 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.650773 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.650837 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs\") pod \"network-metrics-daemon-l47ql\" (UID: \"27d4c184-5dd3-492a-b927-f7a7f31291ef\") " pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:15 crc kubenswrapper[4813]: E0317 09:12:15.650893 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 09:12:15 crc kubenswrapper[4813]: E0317 09:12:15.650943 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 09:12:15 crc kubenswrapper[4813]: E0317 09:12:15.650970 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 09:12:47.650948101 +0000 UTC m=+189.751751610 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 09:12:15 crc kubenswrapper[4813]: E0317 09:12:15.650972 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 09:12:15 crc kubenswrapper[4813]: E0317 09:12:15.650994 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:12:15 crc kubenswrapper[4813]: E0317 09:12:15.651063 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-17 09:12:47.651038034 +0000 UTC m=+189.751841613 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:12:15 crc kubenswrapper[4813]: E0317 09:12:15.651065 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 09:12:15 crc kubenswrapper[4813]: E0317 09:12:15.651124 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 09:12:15 crc kubenswrapper[4813]: E0317 09:12:15.651162 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs podName:27d4c184-5dd3-492a-b927-f7a7f31291ef nodeName:}" failed. No retries permitted until 2026-03-17 09:12:47.651130547 +0000 UTC m=+189.751934106 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs") pod "network-metrics-daemon-l47ql" (UID: "27d4c184-5dd3-492a-b927-f7a7f31291ef") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 09:12:15 crc kubenswrapper[4813]: E0317 09:12:15.651170 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 09:12:15 crc kubenswrapper[4813]: E0317 09:12:15.651202 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:12:15 crc kubenswrapper[4813]: E0317 09:12:15.651238 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 09:12:15 crc kubenswrapper[4813]: E0317 09:12:15.651282 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-17 09:12:47.651258432 +0000 UTC m=+189.752061971 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.650888 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:15 crc kubenswrapper[4813]: E0317 09:12:15.651338 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 09:12:47.651308654 +0000 UTC m=+189.752112193 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.665194 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:15Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.678576 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:15Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.693308 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:15Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.715059 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238868fbd0e0229eec91acd7a54dcd22c659471003620351cba19de858aef8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:15Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.730376 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:15 crc kubenswrapper[4813]: E0317 09:12:15.730569 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.730368 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.730390 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:15 crc kubenswrapper[4813]: E0317 09:12:15.730846 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:15 crc kubenswrapper[4813]: E0317 09:12:15.730954 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.731744 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:15Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.750833 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:15Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.769197 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ecb2c81-a49d-4d2e-83b2-b821ea50c5e5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80403c9745d70069a406ca4771940c9ff6958417787d62595bf005db27d2c1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a5b5d0590fefa483bcb5d5a78dfee7d61881fe9e19e5f229d3788d2cc09ab87\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:10:07Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0317 09:09:41.118018 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0317 09:09:41.120994 1 observer_polling.go:159] Starting file observer\\\\nI0317 09:09:41.159745 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0317 09:09:41.162501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0317 09:10:07.008700 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0317 09:10:07.008843 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:06Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1d7eaebd9e22a3f68b00c57f0bc9971826fc05a41062947a1ce04f1e8d380c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa97f1c57fb3a119266cee961a8872c9115b7541542c57598511fd47ea3c907\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3147869c6b9a5426228139c412d8b7b2cb224a771ca092fbe9fe3b9b2b215d61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:15Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.787730 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:15Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.806491 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:15Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.827699 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:15Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:15 crc kubenswrapper[4813]: I0317 09:12:15.859106 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:13Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 09:12:13.643423 7150 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0317 09:12:13.643461 7150 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0317 09:12:13.643501 7150 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 09:12:13.643512 7150 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 09:12:13.643511 7150 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 09:12:13.643528 7150 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 09:12:13.643536 7150 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 09:12:13.643582 7150 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 09:12:13.643623 7150 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 09:12:13.643624 7150 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0317 09:12:13.643639 7150 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 09:12:13.643659 7150 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0317 09:12:13.643684 7150 factory.go:656] Stopping watch factory\\\\nI0317 09:12:13.643691 7150 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0317 09:12:13.643701 7150 ovnkube.go:599] Stopped ovnkube\\\\nI0317 09:12:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:12:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lngs6_openshift-ovn-kubernetes(a1730ca2-a1bb-4e9a-ba56-7edaded79f36)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:15Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:16 crc kubenswrapper[4813]: I0317 09:12:16.730247 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:16 crc kubenswrapper[4813]: E0317 09:12:16.730397 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:17 crc kubenswrapper[4813]: I0317 09:12:17.730544 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:17 crc kubenswrapper[4813]: I0317 09:12:17.730590 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:17 crc kubenswrapper[4813]: I0317 09:12:17.730646 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:17 crc kubenswrapper[4813]: E0317 09:12:17.730810 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:17 crc kubenswrapper[4813]: E0317 09:12:17.731084 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:17 crc kubenswrapper[4813]: E0317 09:12:17.731475 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.730737 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:18 crc kubenswrapper[4813]: E0317 09:12:18.730862 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.731349 4813 scope.go:117] "RemoveContainer" containerID="05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f" Mar 17 09:12:18 crc kubenswrapper[4813]: E0317 09:12:18.731499 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.750252 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:18Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.768206 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:18Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.790929 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238868fbd0e0229eec91acd7a54dcd22c659471003620351cba19de858aef8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:18Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.807453 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:18Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.825808 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:18Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.844759 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ecb2c81-a49d-4d2e-83b2-b821ea50c5e5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80403c9745d70069a406ca4771940c9ff6958417787d62595bf005db27d2c1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a5b5d0590fefa483bcb5d5a78dfee7d61881fe9e19e5f229d3788d2cc09ab87\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:10:07Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0317 09:09:41.118018 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0317 09:09:41.120994 1 observer_polling.go:159] Starting file observer\\\\nI0317 09:09:41.159745 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0317 09:09:41.162501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0317 09:10:07.008700 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0317 09:10:07.008843 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:06Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1d7eaebd9e22a3f68b00c57f0bc9971826fc05a41062947a1ce04f1e8d380c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa97f1c57fb3a119266cee961a8872c9115b7541542c57598511fd47ea3c907\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3147869c6b9a5426228139c412d8b7b2cb224a771ca092fbe9fe3b9b2b215d61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:18Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.858549 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:18Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:18 crc kubenswrapper[4813]: E0317 09:12:18.875253 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.878815 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:18Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.893379 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:18Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.910700 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.910726 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.910735 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.910749 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.910759 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:18Z","lastTransitionTime":"2026-03-17T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.914011 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:13Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 09:12:13.643423 7150 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0317 09:12:13.643461 7150 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0317 09:12:13.643501 7150 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 09:12:13.643512 7150 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 09:12:13.643511 7150 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 09:12:13.643528 7150 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 09:12:13.643536 7150 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 09:12:13.643582 7150 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 09:12:13.643623 7150 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 09:12:13.643624 7150 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0317 09:12:13.643639 7150 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 09:12:13.643659 7150 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0317 09:12:13.643684 7150 factory.go:656] Stopping watch factory\\\\nI0317 09:12:13.643691 7150 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0317 09:12:13.643701 7150 ovnkube.go:599] Stopped ovnkube\\\\nI0317 09:12:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:12:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lngs6_openshift-ovn-kubernetes(a1730ca2-a1bb-4e9a-ba56-7edaded79f36)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:18Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:18 crc kubenswrapper[4813]: E0317 09:12:18.929669 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:18Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.933514 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.933553 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.933564 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.933580 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.933590 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:18Z","lastTransitionTime":"2026-03-17T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.935957 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:18Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:18 crc kubenswrapper[4813]: E0317 09:12:18.947527 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:18Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.950920 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.950951 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.950963 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.950980 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.950992 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:18Z","lastTransitionTime":"2026-03-17T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.952934 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:18Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.964395 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:18Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:18 crc kubenswrapper[4813]: E0317 09:12:18.965991 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:18Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.969741 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.969800 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.969825 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.969942 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.969975 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:18Z","lastTransitionTime":"2026-03-17T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.980126 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:18Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:18 crc kubenswrapper[4813]: E0317 09:12:18.987930 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:18Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.990881 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:18Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.991056 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.991107 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.991120 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.991139 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:18 crc kubenswrapper[4813]: I0317 09:12:18.991151 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:18Z","lastTransitionTime":"2026-03-17T09:12:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:19 crc kubenswrapper[4813]: I0317 09:12:19.000418 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:18Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:19 crc kubenswrapper[4813]: E0317 09:12:19.002061 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:18Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:19Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:19 crc kubenswrapper[4813]: E0317 09:12:19.002168 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 17 09:12:19 crc kubenswrapper[4813]: I0317 09:12:19.010637 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:19Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:19 crc kubenswrapper[4813]: I0317 09:12:19.729850 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:19 crc kubenswrapper[4813]: I0317 09:12:19.729966 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:19 crc kubenswrapper[4813]: E0317 09:12:19.730075 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:19 crc kubenswrapper[4813]: I0317 09:12:19.729850 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:19 crc kubenswrapper[4813]: E0317 09:12:19.730239 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:19 crc kubenswrapper[4813]: E0317 09:12:19.730407 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:20 crc kubenswrapper[4813]: I0317 09:12:20.729733 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:20 crc kubenswrapper[4813]: E0317 09:12:20.729915 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:21 crc kubenswrapper[4813]: I0317 09:12:21.730405 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:21 crc kubenswrapper[4813]: I0317 09:12:21.730466 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:21 crc kubenswrapper[4813]: I0317 09:12:21.730474 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:21 crc kubenswrapper[4813]: E0317 09:12:21.730635 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:21 crc kubenswrapper[4813]: E0317 09:12:21.730818 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:21 crc kubenswrapper[4813]: E0317 09:12:21.731014 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:22 crc kubenswrapper[4813]: I0317 09:12:22.730564 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:22 crc kubenswrapper[4813]: E0317 09:12:22.730781 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:23 crc kubenswrapper[4813]: I0317 09:12:23.730169 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:23 crc kubenswrapper[4813]: I0317 09:12:23.730232 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:23 crc kubenswrapper[4813]: I0317 09:12:23.730276 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:23 crc kubenswrapper[4813]: E0317 09:12:23.730358 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:23 crc kubenswrapper[4813]: E0317 09:12:23.730638 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:23 crc kubenswrapper[4813]: E0317 09:12:23.730787 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:23 crc kubenswrapper[4813]: E0317 09:12:23.876364 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:12:24 crc kubenswrapper[4813]: I0317 09:12:24.730239 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:24 crc kubenswrapper[4813]: E0317 09:12:24.730583 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:24 crc kubenswrapper[4813]: I0317 09:12:24.743305 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 17 09:12:25 crc kubenswrapper[4813]: I0317 09:12:25.729769 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:25 crc kubenswrapper[4813]: I0317 09:12:25.729866 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:25 crc kubenswrapper[4813]: I0317 09:12:25.730035 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:25 crc kubenswrapper[4813]: E0317 09:12:25.730878 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:25 crc kubenswrapper[4813]: E0317 09:12:25.730976 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:25 crc kubenswrapper[4813]: E0317 09:12:25.731005 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:25 crc kubenswrapper[4813]: I0317 09:12:25.731264 4813 scope.go:117] "RemoveContainer" containerID="1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f" Mar 17 09:12:25 crc kubenswrapper[4813]: E0317 09:12:25.731549 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lngs6_openshift-ovn-kubernetes(a1730ca2-a1bb-4e9a-ba56-7edaded79f36)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" Mar 17 09:12:26 crc kubenswrapper[4813]: I0317 09:12:26.730280 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:26 crc kubenswrapper[4813]: E0317 09:12:26.730535 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:27 crc kubenswrapper[4813]: I0317 09:12:27.730053 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:27 crc kubenswrapper[4813]: I0317 09:12:27.730082 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:27 crc kubenswrapper[4813]: I0317 09:12:27.730077 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:27 crc kubenswrapper[4813]: E0317 09:12:27.730246 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:27 crc kubenswrapper[4813]: E0317 09:12:27.730383 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:27 crc kubenswrapper[4813]: E0317 09:12:27.730520 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:28 crc kubenswrapper[4813]: I0317 09:12:28.730180 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:28 crc kubenswrapper[4813]: E0317 09:12:28.730304 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:28 crc kubenswrapper[4813]: I0317 09:12:28.746824 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:28Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:28 crc kubenswrapper[4813]: I0317 09:12:28.759457 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fabf428b-a70a-4445-aea5-e23e9e6c5dff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://342e8bf02207ce28830cf7d394c4193c7c6722454114ba9febcf14cdd62d7ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03f72a274657c1a544b773168b70bde00e45c5eae7dde99271f43ce5ad87db72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03f72a274657c1a544b773168b70bde00e45c5eae7dde99271f43ce5ad87db72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:28Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:28 crc kubenswrapper[4813]: I0317 09:12:28.774408 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:28Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:28 crc kubenswrapper[4813]: I0317 09:12:28.789824 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:28Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:28 crc kubenswrapper[4813]: I0317 09:12:28.804129 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:28Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:28 crc kubenswrapper[4813]: I0317 09:12:28.821277 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:28Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:28 crc kubenswrapper[4813]: I0317 09:12:28.838102 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:28Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:28 crc kubenswrapper[4813]: I0317 09:12:28.853414 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238868fbd0e0229eec91acd7a54dcd22c659471003620351cba19de858aef8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:28Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:28 crc kubenswrapper[4813]: I0317 09:12:28.869998 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:28Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:28 crc kubenswrapper[4813]: E0317 09:12:28.877248 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:12:28 crc kubenswrapper[4813]: I0317 09:12:28.891295 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:13Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 09:12:13.643423 7150 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0317 09:12:13.643461 7150 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0317 09:12:13.643501 7150 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 09:12:13.643512 7150 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 09:12:13.643511 7150 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 09:12:13.643528 7150 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 09:12:13.643536 7150 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 09:12:13.643582 7150 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 09:12:13.643623 7150 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 09:12:13.643624 7150 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0317 09:12:13.643639 7150 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 09:12:13.643659 7150 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0317 09:12:13.643684 7150 factory.go:656] Stopping watch factory\\\\nI0317 09:12:13.643691 7150 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0317 09:12:13.643701 7150 ovnkube.go:599] Stopped ovnkube\\\\nI0317 09:12:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:12:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lngs6_openshift-ovn-kubernetes(a1730ca2-a1bb-4e9a-ba56-7edaded79f36)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:28Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:28 crc kubenswrapper[4813]: I0317 09:12:28.911680 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ecb2c81-a49d-4d2e-83b2-b821ea50c5e5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80403c9745d70069a406ca4771940c9ff6958417787d62595bf005db27d2c1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a5b5d0590fefa483bcb5d5a78dfee7d61881fe9e19e5f229d3788d2cc09ab87\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:10:07Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0317 09:09:41.118018 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0317 09:09:41.120994 1 observer_polling.go:159] Starting file observer\\\\nI0317 09:09:41.159745 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0317 09:09:41.162501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0317 09:10:07.008700 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0317 09:10:07.008843 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:06Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1d7eaebd9e22a3f68b00c57f0bc9971826fc05a41062947a1ce04f1e8d380c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa97f1c57fb3a119266cee961a8872c9115b7541542c57598511fd47ea3c907\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3147869c6b9a5426228139c412d8b7b2cb224a771ca092fbe9fe3b9b2b215d61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:28Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:28 crc kubenswrapper[4813]: I0317 09:12:28.929291 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:28Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:28 crc kubenswrapper[4813]: I0317 09:12:28.946292 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:28Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:28 crc kubenswrapper[4813]: I0317 09:12:28.964853 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:28Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:28 crc kubenswrapper[4813]: I0317 09:12:28.988311 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:28Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.010098 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:29Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.028547 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:29Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.044392 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:29Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.372351 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.372411 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.372430 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.372454 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.372472 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:29Z","lastTransitionTime":"2026-03-17T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:29 crc kubenswrapper[4813]: E0317 09:12:29.393122 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:29Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.399175 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.399234 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.399251 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.399277 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.399297 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:29Z","lastTransitionTime":"2026-03-17T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:29 crc kubenswrapper[4813]: E0317 09:12:29.421749 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:29Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.427753 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.427802 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.427813 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.427833 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.427850 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:29Z","lastTransitionTime":"2026-03-17T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:29 crc kubenswrapper[4813]: E0317 09:12:29.445323 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:29Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.451224 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.451287 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.451307 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.451332 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.451349 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:29Z","lastTransitionTime":"2026-03-17T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:29 crc kubenswrapper[4813]: E0317 09:12:29.470761 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:29Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.475479 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.475534 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.475554 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.475576 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.475623 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:29Z","lastTransitionTime":"2026-03-17T09:12:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:29 crc kubenswrapper[4813]: E0317 09:12:29.491856 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:29Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:29 crc kubenswrapper[4813]: E0317 09:12:29.492110 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.730216 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.730257 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:29 crc kubenswrapper[4813]: E0317 09:12:29.730418 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:29 crc kubenswrapper[4813]: I0317 09:12:29.730646 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:29 crc kubenswrapper[4813]: E0317 09:12:29.730797 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:29 crc kubenswrapper[4813]: E0317 09:12:29.730888 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:30 crc kubenswrapper[4813]: I0317 09:12:30.729849 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:30 crc kubenswrapper[4813]: E0317 09:12:30.731019 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.598248 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5jjhb_c1246d4d-93d5-4a97-bef8-1ed881e1a217/kube-multus/0.log" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.598304 4813 generic.go:334] "Generic (PLEG): container finished" podID="c1246d4d-93d5-4a97-bef8-1ed881e1a217" containerID="ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737" exitCode=1 Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.598337 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5jjhb" event={"ID":"c1246d4d-93d5-4a97-bef8-1ed881e1a217","Type":"ContainerDied","Data":"ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737"} Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.598772 4813 scope.go:117] "RemoveContainer" containerID="ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.623710 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:31Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.643436 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:31Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.658245 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:31Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.674561 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:31Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.693953 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fabf428b-a70a-4445-aea5-e23e9e6c5dff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://342e8bf02207ce28830cf7d394c4193c7c6722454114ba9febcf14cdd62d7ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03f72a274657c1a544b773168b70bde00e45c5eae7dde99271f43ce5ad87db72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03f72a274657c1a544b773168b70bde00e45c5eae7dde99271f43ce5ad87db72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:31Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.711447 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:31Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.726534 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:31Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.730120 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.730187 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.730132 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:31 crc kubenswrapper[4813]: E0317 09:12:31.730254 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:31 crc kubenswrapper[4813]: E0317 09:12:31.730345 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:31 crc kubenswrapper[4813]: E0317 09:12:31.730492 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.738536 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:31Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.755989 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:31Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.776956 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:31Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.802642 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238868fbd0e0229eec91acd7a54dcd22c659471003620351cba19de858aef8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:31Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.821459 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:31Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.840842 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:31Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:31Z\\\",\\\"message\\\":\\\"2026-03-17T09:11:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_40b34767-c951-4dc6-a04f-ed5954582728\\\\n2026-03-17T09:11:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_40b34767-c951-4dc6-a04f-ed5954582728 to /host/opt/cni/bin/\\\\n2026-03-17T09:11:46Z [verbose] multus-daemon started\\\\n2026-03-17T09:11:46Z [verbose] Readiness Indicator file check\\\\n2026-03-17T09:12:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:31Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.862181 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ecb2c81-a49d-4d2e-83b2-b821ea50c5e5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80403c9745d70069a406ca4771940c9ff6958417787d62595bf005db27d2c1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a5b5d0590fefa483bcb5d5a78dfee7d61881fe9e19e5f229d3788d2cc09ab87\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:10:07Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0317 09:09:41.118018 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0317 09:09:41.120994 1 observer_polling.go:159] Starting file observer\\\\nI0317 09:09:41.159745 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0317 09:09:41.162501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0317 09:10:07.008700 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0317 09:10:07.008843 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:06Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1d7eaebd9e22a3f68b00c57f0bc9971826fc05a41062947a1ce04f1e8d380c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa97f1c57fb3a119266cee961a8872c9115b7541542c57598511fd47ea3c907\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3147869c6b9a5426228139c412d8b7b2cb224a771ca092fbe9fe3b9b2b215d61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:31Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.881817 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:31Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.901485 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:31Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.919507 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:31Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:31 crc kubenswrapper[4813]: I0317 09:12:31.950581 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:13Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 09:12:13.643423 7150 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0317 09:12:13.643461 7150 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0317 09:12:13.643501 7150 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 09:12:13.643512 7150 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 09:12:13.643511 7150 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 09:12:13.643528 7150 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 09:12:13.643536 7150 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 09:12:13.643582 7150 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 09:12:13.643623 7150 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 09:12:13.643624 7150 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0317 09:12:13.643639 7150 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 09:12:13.643659 7150 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0317 09:12:13.643684 7150 factory.go:656] Stopping watch factory\\\\nI0317 09:12:13.643691 7150 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0317 09:12:13.643701 7150 ovnkube.go:599] Stopped ovnkube\\\\nI0317 09:12:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:12:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lngs6_openshift-ovn-kubernetes(a1730ca2-a1bb-4e9a-ba56-7edaded79f36)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:31Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.604789 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5jjhb_c1246d4d-93d5-4a97-bef8-1ed881e1a217/kube-multus/0.log" Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.604874 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5jjhb" event={"ID":"c1246d4d-93d5-4a97-bef8-1ed881e1a217","Type":"ContainerStarted","Data":"3cf6e87ac86b646f6e7e520f8dccf5788696ddef35970778f9490f817a04385a"} Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.623899 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fabf428b-a70a-4445-aea5-e23e9e6c5dff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://342e8bf02207ce28830cf7d394c4193c7c6722454114ba9febcf14cdd62d7ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03f72a274657c1a544b773168b70bde00e45c5eae7dde99271f43ce5ad87db72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03f72a274657c1a544b773168b70bde00e45c5eae7dde99271f43ce5ad87db72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:32Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.642145 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:32Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.659268 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:32Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.675371 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:32Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.695683 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:32Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.713796 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:32Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.730205 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:32 crc kubenswrapper[4813]: E0317 09:12:32.730637 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.730911 4813 scope.go:117] "RemoveContainer" containerID="05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f" Mar 17 09:12:32 crc kubenswrapper[4813]: E0317 09:12:32.731200 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.736056 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238868fbd0e0229eec91acd7a54dcd22c659471003620351cba19de858aef8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:32Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.756356 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:32Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.776515 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cf6e87ac86b646f6e7e520f8dccf5788696ddef35970778f9490f817a04385a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:31Z\\\",\\\"message\\\":\\\"2026-03-17T09:11:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_40b34767-c951-4dc6-a04f-ed5954582728\\\\n2026-03-17T09:11:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_40b34767-c951-4dc6-a04f-ed5954582728 to /host/opt/cni/bin/\\\\n2026-03-17T09:11:46Z [verbose] multus-daemon started\\\\n2026-03-17T09:11:46Z [verbose] Readiness Indicator file check\\\\n2026-03-17T09:12:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:12:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:32Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.795871 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ecb2c81-a49d-4d2e-83b2-b821ea50c5e5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80403c9745d70069a406ca4771940c9ff6958417787d62595bf005db27d2c1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a5b5d0590fefa483bcb5d5a78dfee7d61881fe9e19e5f229d3788d2cc09ab87\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:10:07Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0317 09:09:41.118018 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0317 09:09:41.120994 1 observer_polling.go:159] Starting file observer\\\\nI0317 09:09:41.159745 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0317 09:09:41.162501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0317 09:10:07.008700 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0317 09:10:07.008843 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:06Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1d7eaebd9e22a3f68b00c57f0bc9971826fc05a41062947a1ce04f1e8d380c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa97f1c57fb3a119266cee961a8872c9115b7541542c57598511fd47ea3c907\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3147869c6b9a5426228139c412d8b7b2cb224a771ca092fbe9fe3b9b2b215d61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:32Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.812132 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:32Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.830046 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:32Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.847543 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:32Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.873255 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:13Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 09:12:13.643423 7150 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0317 09:12:13.643461 7150 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0317 09:12:13.643501 7150 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 09:12:13.643512 7150 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 09:12:13.643511 7150 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 09:12:13.643528 7150 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 09:12:13.643536 7150 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 09:12:13.643582 7150 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 09:12:13.643623 7150 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 09:12:13.643624 7150 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0317 09:12:13.643639 7150 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 09:12:13.643659 7150 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0317 09:12:13.643684 7150 factory.go:656] Stopping watch factory\\\\nI0317 09:12:13.643691 7150 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0317 09:12:13.643701 7150 ovnkube.go:599] Stopped ovnkube\\\\nI0317 09:12:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:12:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lngs6_openshift-ovn-kubernetes(a1730ca2-a1bb-4e9a-ba56-7edaded79f36)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:32Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.892571 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:32Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.912297 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:32Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.930824 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:32Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:32 crc kubenswrapper[4813]: I0317 09:12:32.946113 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:32Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:33 crc kubenswrapper[4813]: I0317 09:12:33.730590 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:33 crc kubenswrapper[4813]: I0317 09:12:33.730659 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:33 crc kubenswrapper[4813]: I0317 09:12:33.730638 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:33 crc kubenswrapper[4813]: E0317 09:12:33.730822 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:33 crc kubenswrapper[4813]: E0317 09:12:33.730951 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:33 crc kubenswrapper[4813]: E0317 09:12:33.731169 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:33 crc kubenswrapper[4813]: E0317 09:12:33.878498 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:12:34 crc kubenswrapper[4813]: I0317 09:12:34.730383 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:34 crc kubenswrapper[4813]: E0317 09:12:34.730941 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:35 crc kubenswrapper[4813]: I0317 09:12:35.730151 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:35 crc kubenswrapper[4813]: I0317 09:12:35.730225 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:35 crc kubenswrapper[4813]: E0317 09:12:35.730356 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:35 crc kubenswrapper[4813]: I0317 09:12:35.730170 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:35 crc kubenswrapper[4813]: E0317 09:12:35.730522 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:35 crc kubenswrapper[4813]: E0317 09:12:35.730667 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:36 crc kubenswrapper[4813]: I0317 09:12:36.730190 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:36 crc kubenswrapper[4813]: E0317 09:12:36.730384 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:36 crc kubenswrapper[4813]: I0317 09:12:36.731475 4813 scope.go:117] "RemoveContainer" containerID="1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.626846 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lngs6_a1730ca2-a1bb-4e9a-ba56-7edaded79f36/ovnkube-controller/2.log" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.629740 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerStarted","Data":"09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b"} Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.630712 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.648313 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.661839 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.674509 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.687411 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.699453 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fabf428b-a70a-4445-aea5-e23e9e6c5dff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://342e8bf02207ce28830cf7d394c4193c7c6722454114ba9febcf14cdd62d7ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03f72a274657c1a544b773168b70bde00e45c5eae7dde99271f43ce5ad87db72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03f72a274657c1a544b773168b70bde00e45c5eae7dde99271f43ce5ad87db72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.712580 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.727028 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.729626 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.729688 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:37 crc kubenswrapper[4813]: E0317 09:12:37.729763 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.729693 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:37 crc kubenswrapper[4813]: E0317 09:12:37.729870 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:37 crc kubenswrapper[4813]: E0317 09:12:37.729950 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.738677 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.752430 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.767246 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.788104 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238868fbd0e0229eec91acd7a54dcd22c659471003620351cba19de858aef8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.809384 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.822142 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cf6e87ac86b646f6e7e520f8dccf5788696ddef35970778f9490f817a04385a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:31Z\\\",\\\"message\\\":\\\"2026-03-17T09:11:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_40b34767-c951-4dc6-a04f-ed5954582728\\\\n2026-03-17T09:11:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_40b34767-c951-4dc6-a04f-ed5954582728 to /host/opt/cni/bin/\\\\n2026-03-17T09:11:46Z [verbose] multus-daemon started\\\\n2026-03-17T09:11:46Z [verbose] Readiness Indicator file check\\\\n2026-03-17T09:12:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:12:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.837631 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ecb2c81-a49d-4d2e-83b2-b821ea50c5e5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80403c9745d70069a406ca4771940c9ff6958417787d62595bf005db27d2c1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a5b5d0590fefa483bcb5d5a78dfee7d61881fe9e19e5f229d3788d2cc09ab87\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:10:07Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0317 09:09:41.118018 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0317 09:09:41.120994 1 observer_polling.go:159] Starting file observer\\\\nI0317 09:09:41.159745 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0317 09:09:41.162501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0317 09:10:07.008700 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0317 09:10:07.008843 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:06Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1d7eaebd9e22a3f68b00c57f0bc9971826fc05a41062947a1ce04f1e8d380c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa97f1c57fb3a119266cee961a8872c9115b7541542c57598511fd47ea3c907\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3147869c6b9a5426228139c412d8b7b2cb224a771ca092fbe9fe3b9b2b215d61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.850055 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.860649 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.873880 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:37 crc kubenswrapper[4813]: I0317 09:12:37.891361 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:13Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 09:12:13.643423 7150 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0317 09:12:13.643461 7150 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0317 09:12:13.643501 7150 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 09:12:13.643512 7150 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 09:12:13.643511 7150 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 09:12:13.643528 7150 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 09:12:13.643536 7150 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 09:12:13.643582 7150 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 09:12:13.643623 7150 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 09:12:13.643624 7150 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0317 09:12:13.643639 7150 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 09:12:13.643659 7150 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0317 09:12:13.643684 7150 factory.go:656] Stopping watch factory\\\\nI0317 09:12:13.643691 7150 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0317 09:12:13.643701 7150 ovnkube.go:599] Stopped ovnkube\\\\nI0317 09:12:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:12:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:12:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.637080 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lngs6_a1730ca2-a1bb-4e9a-ba56-7edaded79f36/ovnkube-controller/3.log" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.638048 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lngs6_a1730ca2-a1bb-4e9a-ba56-7edaded79f36/ovnkube-controller/2.log" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.642543 4813 generic.go:334] "Generic (PLEG): container finished" podID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerID="09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b" exitCode=1 Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.642641 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerDied","Data":"09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b"} Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.642734 4813 scope.go:117] "RemoveContainer" containerID="1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.643825 4813 scope.go:117] "RemoveContainer" containerID="09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b" Mar 17 09:12:38 crc kubenswrapper[4813]: E0317 09:12:38.644154 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lngs6_openshift-ovn-kubernetes(a1730ca2-a1bb-4e9a-ba56-7edaded79f36)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.667330 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:38Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.684709 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:38Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.702504 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fabf428b-a70a-4445-aea5-e23e9e6c5dff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://342e8bf02207ce28830cf7d394c4193c7c6722454114ba9febcf14cdd62d7ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03f72a274657c1a544b773168b70bde00e45c5eae7dde99271f43ce5ad87db72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03f72a274657c1a544b773168b70bde00e45c5eae7dde99271f43ce5ad87db72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:38Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.719572 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:38Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.731687 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:38 crc kubenswrapper[4813]: E0317 09:12:38.731835 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.738838 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:38Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.760400 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cf6e87ac86b646f6e7e520f8dccf5788696ddef35970778f9490f817a04385a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:31Z\\\",\\\"message\\\":\\\"2026-03-17T09:11:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_40b34767-c951-4dc6-a04f-ed5954582728\\\\n2026-03-17T09:11:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_40b34767-c951-4dc6-a04f-ed5954582728 to /host/opt/cni/bin/\\\\n2026-03-17T09:11:46Z [verbose] multus-daemon started\\\\n2026-03-17T09:11:46Z [verbose] Readiness Indicator file check\\\\n2026-03-17T09:12:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:12:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:38Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.779875 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:38Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.796235 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:38Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.810122 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238868fbd0e0229eec91acd7a54dcd22c659471003620351cba19de858aef8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:38Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.828268 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:38Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.855980 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:13Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 09:12:13.643423 7150 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0317 09:12:13.643461 7150 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0317 09:12:13.643501 7150 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 09:12:13.643512 7150 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 09:12:13.643511 7150 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 09:12:13.643528 7150 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 09:12:13.643536 7150 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 09:12:13.643582 7150 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 09:12:13.643623 7150 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 09:12:13.643624 7150 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0317 09:12:13.643639 7150 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 09:12:13.643659 7150 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0317 09:12:13.643684 7150 factory.go:656] Stopping watch factory\\\\nI0317 09:12:13.643691 7150 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0317 09:12:13.643701 7150 ovnkube.go:599] Stopped ovnkube\\\\nI0317 09:12:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:12:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:37Z\\\",\\\"message\\\":\\\"ler: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z]\\\\nI0317 09:12:37.795807 7428 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-62stj in node crc\\\\nI0317 09:12:37.794195 7428 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-7xxkf\\\\nI0317 09:12:37.795818 7428 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI0317 09:12:37.795836 7428 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-62stj after 0 failed attempt(s)\\\\nI0317 09:12:37.795860 7428 default_network_controller.go:776] Recording success even\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:12:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:38Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.879054 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ecb2c81-a49d-4d2e-83b2-b821ea50c5e5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80403c9745d70069a406ca4771940c9ff6958417787d62595bf005db27d2c1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a5b5d0590fefa483bcb5d5a78dfee7d61881fe9e19e5f229d3788d2cc09ab87\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:10:07Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0317 09:09:41.118018 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0317 09:09:41.120994 1 observer_polling.go:159] Starting file observer\\\\nI0317 09:09:41.159745 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0317 09:09:41.162501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0317 09:10:07.008700 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0317 09:10:07.008843 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:06Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1d7eaebd9e22a3f68b00c57f0bc9971826fc05a41062947a1ce04f1e8d380c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa97f1c57fb3a119266cee961a8872c9115b7541542c57598511fd47ea3c907\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3147869c6b9a5426228139c412d8b7b2cb224a771ca092fbe9fe3b9b2b215d61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:38Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:38 crc kubenswrapper[4813]: E0317 09:12:38.879791 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.898908 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:38Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.917340 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:38Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.932026 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:38Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.953284 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:38Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.972111 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:38Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:38 crc kubenswrapper[4813]: I0317 09:12:38.991261 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:38Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.008631 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.027316 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cf6e87ac86b646f6e7e520f8dccf5788696ddef35970778f9490f817a04385a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:31Z\\\",\\\"message\\\":\\\"2026-03-17T09:11:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_40b34767-c951-4dc6-a04f-ed5954582728\\\\n2026-03-17T09:11:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_40b34767-c951-4dc6-a04f-ed5954582728 to /host/opt/cni/bin/\\\\n2026-03-17T09:11:46Z [verbose] multus-daemon started\\\\n2026-03-17T09:11:46Z [verbose] Readiness Indicator file check\\\\n2026-03-17T09:12:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:12:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.048129 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.061525 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.083322 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238868fbd0e0229eec91acd7a54dcd22c659471003620351cba19de858aef8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.099971 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.128025 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c479378b211744ef87c8c1a5c301bb35267703a642e6b275d6e1588ca4f354f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:13Z\\\",\\\"message\\\":\\\"e (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0317 09:12:13.643423 7150 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0317 09:12:13.643461 7150 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0317 09:12:13.643501 7150 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0317 09:12:13.643512 7150 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0317 09:12:13.643511 7150 handler.go:208] Removed *v1.Node event handler 7\\\\nI0317 09:12:13.643528 7150 handler.go:208] Removed *v1.Node event handler 2\\\\nI0317 09:12:13.643536 7150 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0317 09:12:13.643582 7150 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0317 09:12:13.643623 7150 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0317 09:12:13.643624 7150 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0317 09:12:13.643639 7150 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0317 09:12:13.643659 7150 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0317 09:12:13.643684 7150 factory.go:656] Stopping watch factory\\\\nI0317 09:12:13.643691 7150 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0317 09:12:13.643701 7150 ovnkube.go:599] Stopped ovnkube\\\\nI0317 09:12:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:12:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:37Z\\\",\\\"message\\\":\\\"ler: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z]\\\\nI0317 09:12:37.795807 7428 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-62stj in node crc\\\\nI0317 09:12:37.794195 7428 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-7xxkf\\\\nI0317 09:12:37.795818 7428 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI0317 09:12:37.795836 7428 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-62stj after 0 failed attempt(s)\\\\nI0317 09:12:37.795860 7428 default_network_controller.go:776] Recording success even\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:12:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.143888 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ecb2c81-a49d-4d2e-83b2-b821ea50c5e5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80403c9745d70069a406ca4771940c9ff6958417787d62595bf005db27d2c1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a5b5d0590fefa483bcb5d5a78dfee7d61881fe9e19e5f229d3788d2cc09ab87\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:10:07Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0317 09:09:41.118018 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0317 09:09:41.120994 1 observer_polling.go:159] Starting file observer\\\\nI0317 09:09:41.159745 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0317 09:09:41.162501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0317 09:10:07.008700 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0317 09:10:07.008843 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:06Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1d7eaebd9e22a3f68b00c57f0bc9971826fc05a41062947a1ce04f1e8d380c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa97f1c57fb3a119266cee961a8872c9115b7541542c57598511fd47ea3c907\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3147869c6b9a5426228139c412d8b7b2cb224a771ca092fbe9fe3b9b2b215d61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.165997 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.179079 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.192728 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.207254 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.219908 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.235675 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.250513 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.263953 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.279569 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fabf428b-a70a-4445-aea5-e23e9e6c5dff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://342e8bf02207ce28830cf7d394c4193c7c6722454114ba9febcf14cdd62d7ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03f72a274657c1a544b773168b70bde00e45c5eae7dde99271f43ce5ad87db72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03f72a274657c1a544b773168b70bde00e45c5eae7dde99271f43ce5ad87db72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.296327 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.648992 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lngs6_a1730ca2-a1bb-4e9a-ba56-7edaded79f36/ovnkube-controller/3.log" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.655289 4813 scope.go:117] "RemoveContainer" containerID="09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b" Mar 17 09:12:39 crc kubenswrapper[4813]: E0317 09:12:39.655659 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lngs6_openshift-ovn-kubernetes(a1730ca2-a1bb-4e9a-ba56-7edaded79f36)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.690488 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:37Z\\\",\\\"message\\\":\\\"ler: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z]\\\\nI0317 09:12:37.795807 7428 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-62stj in node crc\\\\nI0317 09:12:37.794195 7428 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-7xxkf\\\\nI0317 09:12:37.795818 7428 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI0317 09:12:37.795836 7428 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-62stj after 0 failed attempt(s)\\\\nI0317 09:12:37.795860 7428 default_network_controller.go:776] Recording success even\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:12:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lngs6_openshift-ovn-kubernetes(a1730ca2-a1bb-4e9a-ba56-7edaded79f36)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.712569 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ecb2c81-a49d-4d2e-83b2-b821ea50c5e5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80403c9745d70069a406ca4771940c9ff6958417787d62595bf005db27d2c1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a5b5d0590fefa483bcb5d5a78dfee7d61881fe9e19e5f229d3788d2cc09ab87\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:10:07Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0317 09:09:41.118018 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0317 09:09:41.120994 1 observer_polling.go:159] Starting file observer\\\\nI0317 09:09:41.159745 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0317 09:09:41.162501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0317 09:10:07.008700 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0317 09:10:07.008843 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:06Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1d7eaebd9e22a3f68b00c57f0bc9971826fc05a41062947a1ce04f1e8d380c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa97f1c57fb3a119266cee961a8872c9115b7541542c57598511fd47ea3c907\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3147869c6b9a5426228139c412d8b7b2cb224a771ca092fbe9fe3b9b2b215d61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.729751 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.729800 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.729826 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:39 crc kubenswrapper[4813]: E0317 09:12:39.729905 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:39 crc kubenswrapper[4813]: E0317 09:12:39.730090 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:39 crc kubenswrapper[4813]: E0317 09:12:39.730135 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.734726 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.750012 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.750047 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.750054 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.750070 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.750078 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:39Z","lastTransitionTime":"2026-03-17T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.756508 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: E0317 09:12:39.769857 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.776594 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.776711 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.776679 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.776734 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.776886 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.776900 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:39Z","lastTransitionTime":"2026-03-17T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:39 crc kubenswrapper[4813]: E0317 09:12:39.796618 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.800310 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.800969 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.801149 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.801308 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.801507 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.801698 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:39Z","lastTransitionTime":"2026-03-17T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.820382 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: E0317 09:12:39.822416 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.827411 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.827434 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.827443 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.827456 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.827465 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:39Z","lastTransitionTime":"2026-03-17T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.842574 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: E0317 09:12:39.853724 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.857382 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.859065 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.859091 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.859098 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.859114 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.859126 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:39Z","lastTransitionTime":"2026-03-17T09:12:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.871546 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: E0317 09:12:39.882940 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:39Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: E0317 09:12:39.883098 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.886075 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fabf428b-a70a-4445-aea5-e23e9e6c5dff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://342e8bf02207ce28830cf7d394c4193c7c6722454114ba9febcf14cdd62d7ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03f72a274657c1a544b773168b70bde00e45c5eae7dde99271f43ce5ad87db72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03f72a274657c1a544b773168b70bde00e45c5eae7dde99271f43ce5ad87db72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.900159 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.914645 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.931092 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cf6e87ac86b646f6e7e520f8dccf5788696ddef35970778f9490f817a04385a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:31Z\\\",\\\"message\\\":\\\"2026-03-17T09:11:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_40b34767-c951-4dc6-a04f-ed5954582728\\\\n2026-03-17T09:11:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_40b34767-c951-4dc6-a04f-ed5954582728 to /host/opt/cni/bin/\\\\n2026-03-17T09:11:46Z [verbose] multus-daemon started\\\\n2026-03-17T09:11:46Z [verbose] Readiness Indicator file check\\\\n2026-03-17T09:12:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:12:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.944631 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.958187 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.982218 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238868fbd0e0229eec91acd7a54dcd22c659471003620351cba19de858aef8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:39 crc kubenswrapper[4813]: I0317 09:12:39.998659 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:39Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:40 crc kubenswrapper[4813]: I0317 09:12:40.730931 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:40 crc kubenswrapper[4813]: E0317 09:12:40.731098 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:41 crc kubenswrapper[4813]: I0317 09:12:41.730681 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:41 crc kubenswrapper[4813]: I0317 09:12:41.730760 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:41 crc kubenswrapper[4813]: I0317 09:12:41.730805 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:41 crc kubenswrapper[4813]: E0317 09:12:41.730971 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:41 crc kubenswrapper[4813]: E0317 09:12:41.731219 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:41 crc kubenswrapper[4813]: E0317 09:12:41.731285 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:42 crc kubenswrapper[4813]: I0317 09:12:42.730460 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:42 crc kubenswrapper[4813]: E0317 09:12:42.730700 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:43 crc kubenswrapper[4813]: I0317 09:12:43.729927 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:43 crc kubenswrapper[4813]: I0317 09:12:43.729927 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:43 crc kubenswrapper[4813]: I0317 09:12:43.729984 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:43 crc kubenswrapper[4813]: E0317 09:12:43.730640 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:43 crc kubenswrapper[4813]: E0317 09:12:43.730787 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:43 crc kubenswrapper[4813]: E0317 09:12:43.731003 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:43 crc kubenswrapper[4813]: E0317 09:12:43.880994 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:12:44 crc kubenswrapper[4813]: I0317 09:12:44.730411 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:44 crc kubenswrapper[4813]: E0317 09:12:44.730689 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:45 crc kubenswrapper[4813]: I0317 09:12:45.730023 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:45 crc kubenswrapper[4813]: I0317 09:12:45.730016 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:45 crc kubenswrapper[4813]: I0317 09:12:45.730438 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:45 crc kubenswrapper[4813]: E0317 09:12:45.730632 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:45 crc kubenswrapper[4813]: E0317 09:12:45.730660 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:45 crc kubenswrapper[4813]: E0317 09:12:45.730812 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:45 crc kubenswrapper[4813]: I0317 09:12:45.731496 4813 scope.go:117] "RemoveContainer" containerID="05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f" Mar 17 09:12:45 crc kubenswrapper[4813]: E0317 09:12:45.731897 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:12:45 crc kubenswrapper[4813]: I0317 09:12:45.748951 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 17 09:12:46 crc kubenswrapper[4813]: I0317 09:12:46.730814 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:46 crc kubenswrapper[4813]: E0317 09:12:46.732002 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:47 crc kubenswrapper[4813]: I0317 09:12:47.587216 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:12:47 crc kubenswrapper[4813]: E0317 09:12:47.587339 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:51.587317878 +0000 UTC m=+253.688121387 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:12:47 crc kubenswrapper[4813]: I0317 09:12:47.689092 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:47 crc kubenswrapper[4813]: I0317 09:12:47.689840 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs\") pod \"network-metrics-daemon-l47ql\" (UID: \"27d4c184-5dd3-492a-b927-f7a7f31291ef\") " pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:47 crc kubenswrapper[4813]: I0317 09:12:47.689930 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:47 crc kubenswrapper[4813]: I0317 09:12:47.689970 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:47 crc kubenswrapper[4813]: E0317 09:12:47.689284 4813 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 09:12:47 crc kubenswrapper[4813]: E0317 09:12:47.690018 4813 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 09:12:47 crc kubenswrapper[4813]: I0317 09:12:47.690041 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:47 crc kubenswrapper[4813]: E0317 09:12:47.690107 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs podName:27d4c184-5dd3-492a-b927-f7a7f31291ef nodeName:}" failed. No retries permitted until 2026-03-17 09:13:51.690077017 +0000 UTC m=+253.790880546 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs") pod "network-metrics-daemon-l47ql" (UID: "27d4c184-5dd3-492a-b927-f7a7f31291ef") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 17 09:12:47 crc kubenswrapper[4813]: E0317 09:12:47.690133 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 09:13:51.690121499 +0000 UTC m=+253.790925038 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 17 09:12:47 crc kubenswrapper[4813]: E0317 09:12:47.690185 4813 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 09:12:47 crc kubenswrapper[4813]: E0317 09:12:47.690250 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-17 09:13:51.690228062 +0000 UTC m=+253.791031601 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 17 09:12:47 crc kubenswrapper[4813]: E0317 09:12:47.690185 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 09:12:47 crc kubenswrapper[4813]: E0317 09:12:47.690290 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 17 09:12:47 crc kubenswrapper[4813]: E0317 09:12:47.690301 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 09:12:47 crc kubenswrapper[4813]: E0317 09:12:47.690320 4813 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 17 09:12:47 crc kubenswrapper[4813]: E0317 09:12:47.690326 4813 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:12:47 crc kubenswrapper[4813]: E0317 09:12:47.690339 4813 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:12:47 crc kubenswrapper[4813]: E0317 09:12:47.690373 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-17 09:13:51.690360117 +0000 UTC m=+253.791163646 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:12:47 crc kubenswrapper[4813]: E0317 09:12:47.690407 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-17 09:13:51.690384758 +0000 UTC m=+253.791188297 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 17 09:12:47 crc kubenswrapper[4813]: I0317 09:12:47.730150 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:47 crc kubenswrapper[4813]: I0317 09:12:47.730250 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:47 crc kubenswrapper[4813]: E0317 09:12:47.730323 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:47 crc kubenswrapper[4813]: E0317 09:12:47.730438 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:47 crc kubenswrapper[4813]: I0317 09:12:47.730542 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:47 crc kubenswrapper[4813]: E0317 09:12:47.730683 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:48 crc kubenswrapper[4813]: I0317 09:12:48.729913 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:48 crc kubenswrapper[4813]: E0317 09:12:48.730162 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:48 crc kubenswrapper[4813]: I0317 09:12:48.748033 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fabf428b-a70a-4445-aea5-e23e9e6c5dff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://342e8bf02207ce28830cf7d394c4193c7c6722454114ba9febcf14cdd62d7ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03f72a274657c1a544b773168b70bde00e45c5eae7dde99271f43ce5ad87db72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03f72a274657c1a544b773168b70bde00e45c5eae7dde99271f43ce5ad87db72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:48 crc kubenswrapper[4813]: I0317 09:12:48.771419 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2e8434b-0dac-486f-a174-9ec4c9080698\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be5b6cc5bbd218b71befdd26e2f4f660a8e71a5e85caf5e6e9e36f0c3762ca4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b3e4ffb3c9572e4d93aaf170d629b12a806dd4be0eaeb5677ea33e1583b84d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e6d4e31dbb0ec0ee02dd81bd2d654228c3229c7b07de90462c657ee97e45bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e57df849184fd507ce39a1b31b6bef4da5c6c0f9aac69491cadfbc2926e6e0b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbce1d32f7663982ef5c5e5df60dc5a0045f508046991c42a44fad48c7794237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcf270cfbb1e9f689f81743b5f68fd2ad3682e0891da1a27a0727f3c7840eb6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcf270cfbb1e9f689f81743b5f68fd2ad3682e0891da1a27a0727f3c7840eb6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f49c003e0f99daac6c9359b311a0182cd6b638a911503d31bde4c04581f89c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f49c003e0f99daac6c9359b311a0182cd6b638a911503d31bde4c04581f89c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fed1113da1f1a03a32776b473f5e8c699710f4d4b52e0ba795fc5eab447acb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fed1113da1f1a03a32776b473f5e8c699710f4d4b52e0ba795fc5eab447acb49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:48 crc kubenswrapper[4813]: I0317 09:12:48.787897 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:48 crc kubenswrapper[4813]: I0317 09:12:48.801502 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:48 crc kubenswrapper[4813]: I0317 09:12:48.814950 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:48 crc kubenswrapper[4813]: I0317 09:12:48.831532 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:48 crc kubenswrapper[4813]: I0317 09:12:48.843308 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:48 crc kubenswrapper[4813]: I0317 09:12:48.860539 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238868fbd0e0229eec91acd7a54dcd22c659471003620351cba19de858aef8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:48 crc kubenswrapper[4813]: I0317 09:12:48.878627 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:48 crc kubenswrapper[4813]: E0317 09:12:48.881485 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:12:48 crc kubenswrapper[4813]: I0317 09:12:48.892945 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cf6e87ac86b646f6e7e520f8dccf5788696ddef35970778f9490f817a04385a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:31Z\\\",\\\"message\\\":\\\"2026-03-17T09:11:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_40b34767-c951-4dc6-a04f-ed5954582728\\\\n2026-03-17T09:11:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_40b34767-c951-4dc6-a04f-ed5954582728 to /host/opt/cni/bin/\\\\n2026-03-17T09:11:46Z [verbose] multus-daemon started\\\\n2026-03-17T09:11:46Z [verbose] Readiness Indicator file check\\\\n2026-03-17T09:12:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:12:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:48 crc kubenswrapper[4813]: I0317 09:12:48.907519 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ecb2c81-a49d-4d2e-83b2-b821ea50c5e5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80403c9745d70069a406ca4771940c9ff6958417787d62595bf005db27d2c1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a5b5d0590fefa483bcb5d5a78dfee7d61881fe9e19e5f229d3788d2cc09ab87\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:10:07Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0317 09:09:41.118018 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0317 09:09:41.120994 1 observer_polling.go:159] Starting file observer\\\\nI0317 09:09:41.159745 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0317 09:09:41.162501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0317 09:10:07.008700 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0317 09:10:07.008843 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:06Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1d7eaebd9e22a3f68b00c57f0bc9971826fc05a41062947a1ce04f1e8d380c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa97f1c57fb3a119266cee961a8872c9115b7541542c57598511fd47ea3c907\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3147869c6b9a5426228139c412d8b7b2cb224a771ca092fbe9fe3b9b2b215d61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:48 crc kubenswrapper[4813]: I0317 09:12:48.919910 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:48 crc kubenswrapper[4813]: I0317 09:12:48.937363 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:48 crc kubenswrapper[4813]: I0317 09:12:48.952382 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:48 crc kubenswrapper[4813]: I0317 09:12:48.979190 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:37Z\\\",\\\"message\\\":\\\"ler: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z]\\\\nI0317 09:12:37.795807 7428 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-62stj in node crc\\\\nI0317 09:12:37.794195 7428 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-7xxkf\\\\nI0317 09:12:37.795818 7428 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI0317 09:12:37.795836 7428 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-62stj after 0 failed attempt(s)\\\\nI0317 09:12:37.795860 7428 default_network_controller.go:776] Recording success even\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:12:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lngs6_openshift-ovn-kubernetes(a1730ca2-a1bb-4e9a-ba56-7edaded79f36)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:48 crc kubenswrapper[4813]: I0317 09:12:48.997319 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:48Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.011450 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.023127 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.036114 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.730727 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.730771 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:49 crc kubenswrapper[4813]: E0317 09:12:49.730850 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.730785 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:49 crc kubenswrapper[4813]: E0317 09:12:49.731450 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:49 crc kubenswrapper[4813]: E0317 09:12:49.731563 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.925544 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.925584 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.925592 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.925622 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.925632 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:49Z","lastTransitionTime":"2026-03-17T09:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:49 crc kubenswrapper[4813]: E0317 09:12:49.942722 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.947397 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.947428 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.947437 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.947451 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.947461 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:49Z","lastTransitionTime":"2026-03-17T09:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:49 crc kubenswrapper[4813]: E0317 09:12:49.961954 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.966818 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.966865 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.966880 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.966901 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.966916 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:49Z","lastTransitionTime":"2026-03-17T09:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:49 crc kubenswrapper[4813]: E0317 09:12:49.981793 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.986849 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.986949 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.986970 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.986994 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:49 crc kubenswrapper[4813]: I0317 09:12:49.987012 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:49Z","lastTransitionTime":"2026-03-17T09:12:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:50 crc kubenswrapper[4813]: E0317 09:12:50.002166 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:49Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:49Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:50 crc kubenswrapper[4813]: I0317 09:12:50.006831 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:12:50 crc kubenswrapper[4813]: I0317 09:12:50.006894 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:12:50 crc kubenswrapper[4813]: I0317 09:12:50.006915 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:12:50 crc kubenswrapper[4813]: I0317 09:12:50.006942 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:12:50 crc kubenswrapper[4813]: I0317 09:12:50.006964 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:12:50Z","lastTransitionTime":"2026-03-17T09:12:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:12:50 crc kubenswrapper[4813]: E0317 09:12:50.026724 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:12:50Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:50Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:50 crc kubenswrapper[4813]: E0317 09:12:50.026979 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 17 09:12:50 crc kubenswrapper[4813]: I0317 09:12:50.730736 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:50 crc kubenswrapper[4813]: E0317 09:12:50.730918 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:51 crc kubenswrapper[4813]: I0317 09:12:51.729991 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:51 crc kubenswrapper[4813]: I0317 09:12:51.730038 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:51 crc kubenswrapper[4813]: I0317 09:12:51.730062 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:51 crc kubenswrapper[4813]: E0317 09:12:51.730183 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:51 crc kubenswrapper[4813]: E0317 09:12:51.730279 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:51 crc kubenswrapper[4813]: E0317 09:12:51.730395 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:52 crc kubenswrapper[4813]: I0317 09:12:52.729848 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:52 crc kubenswrapper[4813]: E0317 09:12:52.730116 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:53 crc kubenswrapper[4813]: I0317 09:12:53.730591 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:53 crc kubenswrapper[4813]: I0317 09:12:53.730718 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:53 crc kubenswrapper[4813]: I0317 09:12:53.730796 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:53 crc kubenswrapper[4813]: E0317 09:12:53.731038 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:53 crc kubenswrapper[4813]: E0317 09:12:53.731146 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:53 crc kubenswrapper[4813]: E0317 09:12:53.731284 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:53 crc kubenswrapper[4813]: E0317 09:12:53.883298 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:12:54 crc kubenswrapper[4813]: I0317 09:12:54.730334 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:54 crc kubenswrapper[4813]: E0317 09:12:54.730468 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:54 crc kubenswrapper[4813]: I0317 09:12:54.731331 4813 scope.go:117] "RemoveContainer" containerID="09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b" Mar 17 09:12:54 crc kubenswrapper[4813]: E0317 09:12:54.731509 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lngs6_openshift-ovn-kubernetes(a1730ca2-a1bb-4e9a-ba56-7edaded79f36)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" Mar 17 09:12:55 crc kubenswrapper[4813]: I0317 09:12:55.729580 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:55 crc kubenswrapper[4813]: I0317 09:12:55.729686 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:55 crc kubenswrapper[4813]: I0317 09:12:55.729652 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:55 crc kubenswrapper[4813]: E0317 09:12:55.729814 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:55 crc kubenswrapper[4813]: E0317 09:12:55.729990 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:55 crc kubenswrapper[4813]: E0317 09:12:55.730121 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:56 crc kubenswrapper[4813]: I0317 09:12:56.730509 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:56 crc kubenswrapper[4813]: E0317 09:12:56.730710 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:57 crc kubenswrapper[4813]: I0317 09:12:57.729967 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:57 crc kubenswrapper[4813]: I0317 09:12:57.730030 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:57 crc kubenswrapper[4813]: I0317 09:12:57.730075 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:57 crc kubenswrapper[4813]: E0317 09:12:57.731594 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:57 crc kubenswrapper[4813]: E0317 09:12:57.731768 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:57 crc kubenswrapper[4813]: E0317 09:12:57.731865 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:12:58 crc kubenswrapper[4813]: I0317 09:12:58.730562 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:12:58 crc kubenswrapper[4813]: E0317 09:12:58.731069 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:12:58 crc kubenswrapper[4813]: I0317 09:12:58.731320 4813 scope.go:117] "RemoveContainer" containerID="05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f" Mar 17 09:12:58 crc kubenswrapper[4813]: E0317 09:12:58.731555 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 17 09:12:58 crc kubenswrapper[4813]: I0317 09:12:58.751546 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:58 crc kubenswrapper[4813]: I0317 09:12:58.766668 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-62stj" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8a42f832-caac-4d43-aba1-63eab74eb5f5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://845257c4f1a23fc4d0199022ce308c239108e44b121628aff44d5c4ff636b0db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zs4ft\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-62stj\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:58 crc kubenswrapper[4813]: I0317 09:12:58.788514 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a21775d2-fd48-430d-bd5c-679b7bb2c271\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:11:42Z\\\",\\\"message\\\":\\\"le observer\\\\nW0317 09:11:42.514206 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0317 09:11:42.514314 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0317 09:11:42.514975 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2864377217/tls.crt::/tmp/serving-cert-2864377217/tls.key\\\\\\\"\\\\nI0317 09:11:42.789791 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0317 09:11:42.792563 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0317 09:11:42.792620 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0317 09:11:42.792650 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0317 09:11:42.792659 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0317 09:11:42.798228 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW0317 09:11:42.798251 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798256 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0317 09:11:42.798260 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0317 09:11:42.798263 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0317 09:11:42.798265 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0317 09:11:42.798268 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI0317 09:11:42.798282 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF0317 09:11:42.801047 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:41Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 1m20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:58 crc kubenswrapper[4813]: I0317 09:12:58.807848 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:58 crc kubenswrapper[4813]: I0317 09:12:58.825670 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e7e156ce-b279-4a25-a27e-2da7d0ba6437\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://290d30ac175ecabe9ebec9b39be1242ebeb12bf7b5c79c6013d5537ec69988be\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-26dhg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-7xxkf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:58 crc kubenswrapper[4813]: I0317 09:12:58.842250 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-l47ql" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"27d4c184-5dd3-492a-b927-f7a7f31291ef\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wd8rt\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-l47ql\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:58 crc kubenswrapper[4813]: I0317 09:12:58.856768 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-zr9rn" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"95fc4db5-3177-485a-adfd-f829495e487a\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c229f8f9fd15600c5b67bb885d1b8237e188a18ad6567e339f1b0801de677bfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m6hct\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-zr9rn\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:58 crc kubenswrapper[4813]: I0317 09:12:58.872864 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fabf428b-a70a-4445-aea5-e23e9e6c5dff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://342e8bf02207ce28830cf7d394c4193c7c6722454114ba9febcf14cdd62d7ae5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://03f72a274657c1a544b773168b70bde00e45c5eae7dde99271f43ce5ad87db72\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://03f72a274657c1a544b773168b70bde00e45c5eae7dde99271f43ce5ad87db72\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:58 crc kubenswrapper[4813]: E0317 09:12:58.883906 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:12:58 crc kubenswrapper[4813]: I0317 09:12:58.908113 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f2e8434b-0dac-486f-a174-9ec4c9080698\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:42Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://be5b6cc5bbd218b71befdd26e2f4f660a8e71a5e85caf5e6e9e36f0c3762ca4a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4b3e4ffb3c9572e4d93aaf170d629b12a806dd4be0eaeb5677ea33e1583b84d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://4e6d4e31dbb0ec0ee02dd81bd2d654228c3229c7b07de90462c657ee97e45bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://e57df849184fd507ce39a1b31b6bef4da5c6c0f9aac69491cadfbc2926e6e0b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bbce1d32f7663982ef5c5e5df60dc5a0045f508046991c42a44fad48c7794237\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bcf270cfbb1e9f689f81743b5f68fd2ad3682e0891da1a27a0727f3c7840eb6d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcf270cfbb1e9f689f81743b5f68fd2ad3682e0891da1a27a0727f3c7840eb6d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f49c003e0f99daac6c9359b311a0182cd6b638a911503d31bde4c04581f89c99\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f49c003e0f99daac6c9359b311a0182cd6b638a911503d31bde4c04581f89c99\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:41Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://fed1113da1f1a03a32776b473f5e8c699710f4d4b52e0ba795fc5eab447acb49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fed1113da1f1a03a32776b473f5e8c699710f4d4b52e0ba795fc5eab447acb49\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:58 crc kubenswrapper[4813]: I0317 09:12:58.932549 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d35d2ade-2829-4fc2-864d-1871c18006f7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://238868fbd0e0229eec91acd7a54dcd22c659471003620351cba19de858aef8ac\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:50Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5303607c8bfd79128c7e4719ac70bb78d8ba7d128cb8fa7137c5e601ec6c104d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d1f286746970241f9f7f86e1809ede32fc1a9e2dce10b1d9629e8a4092919ac7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998bc4fcaa1d0e6cf7e8eb8d7dee65bf9979b2db79acb826c385cd2375a946d4\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:46Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d870c432205a2614b531a26a2a223d882235da2030b9796a00ddcf530a51146b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f6f2ce27a805c23c52ca09fdec4787919677a11155e14b214cc3b23da15b1eb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ed839730b9aec734ad4f63a18f1bc033391382ae258531cb432fa88af559474f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:49Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wmxwh\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-c2dwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:58 crc kubenswrapper[4813]: I0317 09:12:58.953433 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8ac387df-f9b1-4ce9-a109-1b80e7659f2e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5294bde5f6e92db44dcd0b9db344e8d1d0884bc8ee795bbd9c94931f06c3fd68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ae3a57bb8fb19017c08f4e407c62ef0c00a8a2e327df1abc036caa1ff8d788b2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-xvftk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-bb4gp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:58 crc kubenswrapper[4813]: I0317 09:12:58.974964 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-5jjhb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c1246d4d-93d5-4a97-bef8-1ed881e1a217\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:12:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cf6e87ac86b646f6e7e520f8dccf5788696ddef35970778f9490f817a04385a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:31Z\\\",\\\"message\\\":\\\"2026-03-17T09:11:46+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_40b34767-c951-4dc6-a04f-ed5954582728\\\\n2026-03-17T09:11:46+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_40b34767-c951-4dc6-a04f-ed5954582728 to /host/opt/cni/bin/\\\\n2026-03-17T09:11:46Z [verbose] multus-daemon started\\\\n2026-03-17T09:11:46Z [verbose] Readiness Indicator file check\\\\n2026-03-17T09:12:31Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:12:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-mpt6d\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-multus\"/\"multus-5jjhb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:58 crc kubenswrapper[4813]: I0317 09:12:58.995134 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://19589fbe1806b4129c4544da68175da2e7cad3fa9e6a7755c37ac78fa6219423\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8c48ea0fda34430bdec80582382c7013a9640a664690fe179718f3f99e3d51d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:58Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:59 crc kubenswrapper[4813]: I0317 09:12:59.012892 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:47Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aee8b2ac20904f03c45d3462e6692f47ce04e39f594954a934a87edab2a92ba9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:59Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:59 crc kubenswrapper[4813]: I0317 09:12:59.031121 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:59Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:59 crc kubenswrapper[4813]: I0317 09:12:59.046491 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://09f33554eb4f04c4d79fc3f1ac41e10556a18827dbf9710b6782219250fa541a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:59Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:59 crc kubenswrapper[4813]: I0317 09:12:59.067859 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:45Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:11:43Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:45Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-17T09:12:37Z\\\",\\\"message\\\":\\\"ler: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:37Z is after 2025-08-24T17:21:41Z]\\\\nI0317 09:12:37.795807 7428 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-62stj in node crc\\\\nI0317 09:12:37.794195 7428 default_network_controller.go:776] Recording success event on pod openshift-machine-config-operator/machine-config-daemon-7xxkf\\\\nI0317 09:12:37.795818 7428 obj_retry.go:386] Retry successful for *v1.Pod openshift-network-operator/iptables-alerter-4ln5h after 0 failed attempt(s)\\\\nI0317 09:12:37.795836 7428 obj_retry.go:386] Retry successful for *v1.Pod openshift-image-registry/node-ca-62stj after 0 failed attempt(s)\\\\nI0317 09:12:37.795860 7428 default_network_controller.go:776] Recording success even\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:12:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lngs6_openshift-ovn-kubernetes(a1730ca2-a1bb-4e9a-ba56-7edaded79f36)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:11:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:11:44Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:11:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j4jw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:11:43Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lngs6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:59Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:59 crc kubenswrapper[4813]: I0317 09:12:59.088032 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9ecb2c81-a49d-4d2e-83b2-b821ea50c5e5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80403c9745d70069a406ca4771940c9ff6958417787d62595bf005db27d2c1da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://6a5b5d0590fefa483bcb5d5a78dfee7d61881fe9e19e5f229d3788d2cc09ab87\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-17T09:10:07Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0317 09:09:41.118018 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0317 09:09:41.120994 1 observer_polling.go:159] Starting file observer\\\\nI0317 09:09:41.159745 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0317 09:09:41.162501 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0317 09:10:07.008700 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0317 09:10:07.008843 1 cmd.go:179] failed checking apiserver connectivity: Get \\\\\\\"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/openshift-kube-controller-manager/leases/cluster-policy-controller-lock\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:10:06Z is after 2026-02-23T05:33:13Z\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:10:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ee1d7eaebd9e22a3f68b00c57f0bc9971826fc05a41062947a1ce04f1e8d380c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://faa97f1c57fb3a119266cee961a8872c9115b7541542c57598511fd47ea3c907\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3147869c6b9a5426228139c412d8b7b2cb224a771ca092fbe9fe3b9b2b215d61\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:59Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:59 crc kubenswrapper[4813]: I0317 09:12:59.104954 4813 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a9a16982-ac5f-4b02-a8cc-66b6757bdf0f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:10:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-17T09:09:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a7d3029b3f25cab5a2bc706fe58ec7a9269d638ac8be102cb6ffad85e853162c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb66758e25a18f7add1f3116be245ec337d403192fef407957abf079ec42ae16\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://eb9b14724430a044c8f5a490600a08ced0454b2f1922c0a554b3987880e56ffa\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-17T09:09:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://926456d361c3c7cd7d625021cc2a7cc55636430860ac0dad1523b1e771a92c0f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-17T09:09:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-17T09:09:39Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-17T09:09:38Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:12:59Z is after 2025-08-24T17:21:41Z" Mar 17 09:12:59 crc kubenswrapper[4813]: I0317 09:12:59.730467 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:12:59 crc kubenswrapper[4813]: I0317 09:12:59.730549 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:12:59 crc kubenswrapper[4813]: E0317 09:12:59.730944 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:12:59 crc kubenswrapper[4813]: E0317 09:12:59.731553 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:12:59 crc kubenswrapper[4813]: I0317 09:12:59.731581 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:12:59 crc kubenswrapper[4813]: E0317 09:12:59.731846 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.399777 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.399826 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.399845 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.399868 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.399884 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:13:00Z","lastTransitionTime":"2026-03-17T09:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:13:00 crc kubenswrapper[4813]: E0317 09:13:00.420546 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:13:00Z is after 2025-08-24T17:21:41Z" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.425403 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.425430 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.425440 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.425455 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.425465 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:13:00Z","lastTransitionTime":"2026-03-17T09:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:13:00 crc kubenswrapper[4813]: E0317 09:13:00.443108 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:13:00Z is after 2025-08-24T17:21:41Z" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.447163 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.447230 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.447251 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.447276 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.447296 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:13:00Z","lastTransitionTime":"2026-03-17T09:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:13:00 crc kubenswrapper[4813]: E0317 09:13:00.504292 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:13:00Z is after 2025-08-24T17:21:41Z" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.509269 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.509351 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.509377 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.509406 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.509428 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:13:00Z","lastTransitionTime":"2026-03-17T09:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:13:00 crc kubenswrapper[4813]: E0317 09:13:00.525273 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:13:00Z is after 2025-08-24T17:21:41Z" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.530512 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.530568 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.530588 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.530630 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.530647 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:13:00Z","lastTransitionTime":"2026-03-17T09:13:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:13:00 crc kubenswrapper[4813]: E0317 09:13:00.549065 4813 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-17T09:13:00Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"ea32862a-5b23-419b-aecf-996f0626022b\\\",\\\"systemUUID\\\":\\\"63a6017f-37e1-4eb4-b66c-a7f05e6608e4\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-17T09:13:00Z is after 2025-08-24T17:21:41Z" Mar 17 09:13:00 crc kubenswrapper[4813]: E0317 09:13:00.549428 4813 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 17 09:13:00 crc kubenswrapper[4813]: I0317 09:13:00.730136 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:13:00 crc kubenswrapper[4813]: E0317 09:13:00.730310 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:13:01 crc kubenswrapper[4813]: I0317 09:13:01.729781 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:13:01 crc kubenswrapper[4813]: I0317 09:13:01.729834 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:01 crc kubenswrapper[4813]: I0317 09:13:01.729834 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:01 crc kubenswrapper[4813]: E0317 09:13:01.729997 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:13:01 crc kubenswrapper[4813]: E0317 09:13:01.730288 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:13:01 crc kubenswrapper[4813]: E0317 09:13:01.730457 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:13:02 crc kubenswrapper[4813]: I0317 09:13:02.730172 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:13:02 crc kubenswrapper[4813]: E0317 09:13:02.730351 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:13:03 crc kubenswrapper[4813]: I0317 09:13:03.729790 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:03 crc kubenswrapper[4813]: I0317 09:13:03.729829 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:03 crc kubenswrapper[4813]: I0317 09:13:03.730084 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:13:03 crc kubenswrapper[4813]: E0317 09:13:03.730437 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:13:03 crc kubenswrapper[4813]: E0317 09:13:03.730632 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:13:03 crc kubenswrapper[4813]: E0317 09:13:03.730740 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:13:03 crc kubenswrapper[4813]: E0317 09:13:03.885182 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:13:04 crc kubenswrapper[4813]: I0317 09:13:04.729959 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:13:04 crc kubenswrapper[4813]: E0317 09:13:04.730585 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:13:05 crc kubenswrapper[4813]: I0317 09:13:05.730038 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:05 crc kubenswrapper[4813]: I0317 09:13:05.730065 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:13:05 crc kubenswrapper[4813]: I0317 09:13:05.730103 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:05 crc kubenswrapper[4813]: E0317 09:13:05.730774 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:13:05 crc kubenswrapper[4813]: E0317 09:13:05.730881 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:13:05 crc kubenswrapper[4813]: E0317 09:13:05.730939 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:13:06 crc kubenswrapper[4813]: I0317 09:13:06.730471 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:13:06 crc kubenswrapper[4813]: E0317 09:13:06.730730 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:13:07 crc kubenswrapper[4813]: I0317 09:13:07.730523 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:13:07 crc kubenswrapper[4813]: I0317 09:13:07.730624 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:07 crc kubenswrapper[4813]: I0317 09:13:07.730541 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:07 crc kubenswrapper[4813]: E0317 09:13:07.730759 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:13:07 crc kubenswrapper[4813]: E0317 09:13:07.730873 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:13:07 crc kubenswrapper[4813]: E0317 09:13:07.731062 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:13:08 crc kubenswrapper[4813]: I0317 09:13:08.730909 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:13:08 crc kubenswrapper[4813]: E0317 09:13:08.731092 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:13:08 crc kubenswrapper[4813]: I0317 09:13:08.731504 4813 scope.go:117] "RemoveContainer" containerID="09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b" Mar 17 09:13:08 crc kubenswrapper[4813]: E0317 09:13:08.732782 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lngs6_openshift-ovn-kubernetes(a1730ca2-a1bb-4e9a-ba56-7edaded79f36)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" Mar 17 09:13:08 crc kubenswrapper[4813]: I0317 09:13:08.812695 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-c2dwc" podStartSLOduration=146.812671594 podStartE2EDuration="2m26.812671594s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:08.811679691 +0000 UTC m=+210.912483230" watchObservedRunningTime="2026-03-17 09:13:08.812671594 +0000 UTC m=+210.913475103" Mar 17 09:13:08 crc kubenswrapper[4813]: I0317 09:13:08.830747 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-bb4gp" podStartSLOduration=146.83072898 podStartE2EDuration="2m26.83072898s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:08.830248623 +0000 UTC m=+210.931052162" watchObservedRunningTime="2026-03-17 09:13:08.83072898 +0000 UTC m=+210.931532489" Mar 17 09:13:08 crc kubenswrapper[4813]: E0317 09:13:08.886408 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:13:08 crc kubenswrapper[4813]: I0317 09:13:08.892530 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-5jjhb" podStartSLOduration=146.892497873 podStartE2EDuration="2m26.892497873s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:08.862002795 +0000 UTC m=+210.962806334" watchObservedRunningTime="2026-03-17 09:13:08.892497873 +0000 UTC m=+210.993301412" Mar 17 09:13:08 crc kubenswrapper[4813]: I0317 09:13:08.893751 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=72.893733576 podStartE2EDuration="1m12.893733576s" podCreationTimestamp="2026-03-17 09:11:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:08.893253089 +0000 UTC m=+210.994056658" watchObservedRunningTime="2026-03-17 09:13:08.893733576 +0000 UTC m=+210.994537125" Mar 17 09:13:08 crc kubenswrapper[4813]: I0317 09:13:08.919511 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=79.919486792 podStartE2EDuration="1m19.919486792s" podCreationTimestamp="2026-03-17 09:11:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:08.913033083 +0000 UTC m=+211.013836612" watchObservedRunningTime="2026-03-17 09:13:08.919486792 +0000 UTC m=+211.020290301" Mar 17 09:13:09 crc kubenswrapper[4813]: I0317 09:13:09.051892 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-62stj" podStartSLOduration=147.051875061 podStartE2EDuration="2m27.051875061s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:09.043513676 +0000 UTC m=+211.144317175" watchObservedRunningTime="2026-03-17 09:13:09.051875061 +0000 UTC m=+211.152678560" Mar 17 09:13:09 crc kubenswrapper[4813]: I0317 09:13:09.052468 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=45.052464591 podStartE2EDuration="45.052464591s" podCreationTimestamp="2026-03-17 09:12:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:09.05157286 +0000 UTC m=+211.152376359" watchObservedRunningTime="2026-03-17 09:13:09.052464591 +0000 UTC m=+211.153268090" Mar 17 09:13:09 crc kubenswrapper[4813]: I0317 09:13:09.076682 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=24.076661315 podStartE2EDuration="24.076661315s" podCreationTimestamp="2026-03-17 09:12:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:09.075783145 +0000 UTC m=+211.176586644" watchObservedRunningTime="2026-03-17 09:13:09.076661315 +0000 UTC m=+211.177464824" Mar 17 09:13:09 crc kubenswrapper[4813]: I0317 09:13:09.099559 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podStartSLOduration=147.099533804 podStartE2EDuration="2m27.099533804s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:09.088586081 +0000 UTC m=+211.189389580" watchObservedRunningTime="2026-03-17 09:13:09.099533804 +0000 UTC m=+211.200337313" Mar 17 09:13:09 crc kubenswrapper[4813]: I0317 09:13:09.112341 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-zr9rn" podStartSLOduration=147.112327229 podStartE2EDuration="2m27.112327229s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:09.112212155 +0000 UTC m=+211.213015654" watchObservedRunningTime="2026-03-17 09:13:09.112327229 +0000 UTC m=+211.213130728" Mar 17 09:13:09 crc kubenswrapper[4813]: I0317 09:13:09.730562 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:09 crc kubenswrapper[4813]: I0317 09:13:09.730664 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:09 crc kubenswrapper[4813]: I0317 09:13:09.730563 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:13:09 crc kubenswrapper[4813]: E0317 09:13:09.730855 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:13:09 crc kubenswrapper[4813]: E0317 09:13:09.731052 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:13:09 crc kubenswrapper[4813]: E0317 09:13:09.731283 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.554743 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.554807 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.554826 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.554850 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.554868 4813 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-17T09:13:10Z","lastTransitionTime":"2026-03-17T09:13:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.637119 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-4xfkw"] Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.638225 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4xfkw" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.646386 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.647346 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.647435 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.647986 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.733711 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:13:10 crc kubenswrapper[4813]: E0317 09:13:10.733857 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.761860 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1fb5de2d-d619-4be1-a163-388914906eba-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-4xfkw\" (UID: \"1fb5de2d-d619-4be1-a163-388914906eba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4xfkw" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.761955 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1fb5de2d-d619-4be1-a163-388914906eba-service-ca\") pod \"cluster-version-operator-5c965bbfc6-4xfkw\" (UID: \"1fb5de2d-d619-4be1-a163-388914906eba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4xfkw" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.762012 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1fb5de2d-d619-4be1-a163-388914906eba-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-4xfkw\" (UID: \"1fb5de2d-d619-4be1-a163-388914906eba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4xfkw" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.762033 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fb5de2d-d619-4be1-a163-388914906eba-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-4xfkw\" (UID: \"1fb5de2d-d619-4be1-a163-388914906eba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4xfkw" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.762063 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1fb5de2d-d619-4be1-a163-388914906eba-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-4xfkw\" (UID: \"1fb5de2d-d619-4be1-a163-388914906eba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4xfkw" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.772658 4813 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.780358 4813 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.863690 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1fb5de2d-d619-4be1-a163-388914906eba-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-4xfkw\" (UID: \"1fb5de2d-d619-4be1-a163-388914906eba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4xfkw" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.863765 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fb5de2d-d619-4be1-a163-388914906eba-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-4xfkw\" (UID: \"1fb5de2d-d619-4be1-a163-388914906eba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4xfkw" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.863848 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1fb5de2d-d619-4be1-a163-388914906eba-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-4xfkw\" (UID: \"1fb5de2d-d619-4be1-a163-388914906eba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4xfkw" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.863940 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1fb5de2d-d619-4be1-a163-388914906eba-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-4xfkw\" (UID: \"1fb5de2d-d619-4be1-a163-388914906eba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4xfkw" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.864001 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1fb5de2d-d619-4be1-a163-388914906eba-service-ca\") pod \"cluster-version-operator-5c965bbfc6-4xfkw\" (UID: \"1fb5de2d-d619-4be1-a163-388914906eba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4xfkw" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.864303 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/1fb5de2d-d619-4be1-a163-388914906eba-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-4xfkw\" (UID: \"1fb5de2d-d619-4be1-a163-388914906eba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4xfkw" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.864575 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/1fb5de2d-d619-4be1-a163-388914906eba-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-4xfkw\" (UID: \"1fb5de2d-d619-4be1-a163-388914906eba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4xfkw" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.868264 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/1fb5de2d-d619-4be1-a163-388914906eba-service-ca\") pod \"cluster-version-operator-5c965bbfc6-4xfkw\" (UID: \"1fb5de2d-d619-4be1-a163-388914906eba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4xfkw" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.874525 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1fb5de2d-d619-4be1-a163-388914906eba-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-4xfkw\" (UID: \"1fb5de2d-d619-4be1-a163-388914906eba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4xfkw" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.882486 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1fb5de2d-d619-4be1-a163-388914906eba-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-4xfkw\" (UID: \"1fb5de2d-d619-4be1-a163-388914906eba\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4xfkw" Mar 17 09:13:10 crc kubenswrapper[4813]: I0317 09:13:10.968907 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4xfkw" Mar 17 09:13:11 crc kubenswrapper[4813]: I0317 09:13:11.729935 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:13:11 crc kubenswrapper[4813]: E0317 09:13:11.730513 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:13:11 crc kubenswrapper[4813]: I0317 09:13:11.729987 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:11 crc kubenswrapper[4813]: E0317 09:13:11.730679 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:13:11 crc kubenswrapper[4813]: I0317 09:13:11.729935 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:11 crc kubenswrapper[4813]: E0317 09:13:11.730805 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:13:11 crc kubenswrapper[4813]: I0317 09:13:11.766431 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4xfkw" event={"ID":"1fb5de2d-d619-4be1-a163-388914906eba","Type":"ContainerStarted","Data":"f180ae9222bdff6a04dd7b6e18c3aba6b080ab9a53d75d543bf244e41d3c8dd5"} Mar 17 09:13:11 crc kubenswrapper[4813]: I0317 09:13:11.766502 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4xfkw" event={"ID":"1fb5de2d-d619-4be1-a163-388914906eba","Type":"ContainerStarted","Data":"f67e7612675454496e03454a22103df89ab776355bab4fc94c2f3c0a01fae33e"} Mar 17 09:13:11 crc kubenswrapper[4813]: I0317 09:13:11.788211 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-4xfkw" podStartSLOduration=149.788181346 podStartE2EDuration="2m29.788181346s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:11.785324198 +0000 UTC m=+213.886127727" watchObservedRunningTime="2026-03-17 09:13:11.788181346 +0000 UTC m=+213.888984885" Mar 17 09:13:12 crc kubenswrapper[4813]: I0317 09:13:12.730071 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:13:12 crc kubenswrapper[4813]: E0317 09:13:12.730674 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:13:12 crc kubenswrapper[4813]: I0317 09:13:12.731020 4813 scope.go:117] "RemoveContainer" containerID="05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f" Mar 17 09:13:13 crc kubenswrapper[4813]: I0317 09:13:13.730332 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:13 crc kubenswrapper[4813]: I0317 09:13:13.730332 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:13 crc kubenswrapper[4813]: E0317 09:13:13.730476 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:13:13 crc kubenswrapper[4813]: I0317 09:13:13.730349 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:13:13 crc kubenswrapper[4813]: E0317 09:13:13.730912 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:13:13 crc kubenswrapper[4813]: E0317 09:13:13.731025 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:13:13 crc kubenswrapper[4813]: I0317 09:13:13.775191 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/4.log" Mar 17 09:13:13 crc kubenswrapper[4813]: I0317 09:13:13.777447 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d76b00866bfb10cc5b5d6e623f6ae3e733020c40adf36dd1354bce3bddc64986"} Mar 17 09:13:13 crc kubenswrapper[4813]: I0317 09:13:13.777934 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:13:13 crc kubenswrapper[4813]: E0317 09:13:13.887505 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:13:14 crc kubenswrapper[4813]: I0317 09:13:14.729985 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:13:14 crc kubenswrapper[4813]: E0317 09:13:14.730160 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:13:15 crc kubenswrapper[4813]: I0317 09:13:15.730583 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:13:15 crc kubenswrapper[4813]: I0317 09:13:15.730652 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:15 crc kubenswrapper[4813]: I0317 09:13:15.730660 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:15 crc kubenswrapper[4813]: E0317 09:13:15.730773 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:13:15 crc kubenswrapper[4813]: E0317 09:13:15.730940 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:13:15 crc kubenswrapper[4813]: E0317 09:13:15.731041 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:13:16 crc kubenswrapper[4813]: I0317 09:13:16.730206 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:13:16 crc kubenswrapper[4813]: E0317 09:13:16.730450 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:13:17 crc kubenswrapper[4813]: I0317 09:13:17.730671 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:17 crc kubenswrapper[4813]: I0317 09:13:17.730743 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:17 crc kubenswrapper[4813]: I0317 09:13:17.730743 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:13:17 crc kubenswrapper[4813]: E0317 09:13:17.730870 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:13:17 crc kubenswrapper[4813]: E0317 09:13:17.731094 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:13:17 crc kubenswrapper[4813]: E0317 09:13:17.731172 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:13:17 crc kubenswrapper[4813]: I0317 09:13:17.796828 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5jjhb_c1246d4d-93d5-4a97-bef8-1ed881e1a217/kube-multus/1.log" Mar 17 09:13:17 crc kubenswrapper[4813]: I0317 09:13:17.797794 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5jjhb_c1246d4d-93d5-4a97-bef8-1ed881e1a217/kube-multus/0.log" Mar 17 09:13:17 crc kubenswrapper[4813]: I0317 09:13:17.797894 4813 generic.go:334] "Generic (PLEG): container finished" podID="c1246d4d-93d5-4a97-bef8-1ed881e1a217" containerID="3cf6e87ac86b646f6e7e520f8dccf5788696ddef35970778f9490f817a04385a" exitCode=1 Mar 17 09:13:17 crc kubenswrapper[4813]: I0317 09:13:17.797950 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5jjhb" event={"ID":"c1246d4d-93d5-4a97-bef8-1ed881e1a217","Type":"ContainerDied","Data":"3cf6e87ac86b646f6e7e520f8dccf5788696ddef35970778f9490f817a04385a"} Mar 17 09:13:17 crc kubenswrapper[4813]: I0317 09:13:17.798054 4813 scope.go:117] "RemoveContainer" containerID="ec9a0499d878f1a58bab7c7a7f468c5a586d1e04e094530d019db7f266aa5737" Mar 17 09:13:17 crc kubenswrapper[4813]: I0317 09:13:17.798859 4813 scope.go:117] "RemoveContainer" containerID="3cf6e87ac86b646f6e7e520f8dccf5788696ddef35970778f9490f817a04385a" Mar 17 09:13:17 crc kubenswrapper[4813]: E0317 09:13:17.799153 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-5jjhb_openshift-multus(c1246d4d-93d5-4a97-bef8-1ed881e1a217)\"" pod="openshift-multus/multus-5jjhb" podUID="c1246d4d-93d5-4a97-bef8-1ed881e1a217" Mar 17 09:13:17 crc kubenswrapper[4813]: I0317 09:13:17.831926 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=88.831899913 podStartE2EDuration="1m28.831899913s" podCreationTimestamp="2026-03-17 09:11:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:13.799410577 +0000 UTC m=+215.900214076" watchObservedRunningTime="2026-03-17 09:13:17.831899913 +0000 UTC m=+219.932703442" Mar 17 09:13:18 crc kubenswrapper[4813]: I0317 09:13:18.730124 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:13:18 crc kubenswrapper[4813]: E0317 09:13:18.731525 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:13:18 crc kubenswrapper[4813]: I0317 09:13:18.804091 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5jjhb_c1246d4d-93d5-4a97-bef8-1ed881e1a217/kube-multus/1.log" Mar 17 09:13:18 crc kubenswrapper[4813]: E0317 09:13:18.888531 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:13:19 crc kubenswrapper[4813]: I0317 09:13:19.729873 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:19 crc kubenswrapper[4813]: I0317 09:13:19.730005 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:13:19 crc kubenswrapper[4813]: I0317 09:13:19.730248 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:19 crc kubenswrapper[4813]: E0317 09:13:19.730234 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:13:19 crc kubenswrapper[4813]: E0317 09:13:19.730335 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:13:19 crc kubenswrapper[4813]: E0317 09:13:19.730434 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:13:20 crc kubenswrapper[4813]: I0317 09:13:20.729901 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:13:20 crc kubenswrapper[4813]: E0317 09:13:20.730079 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:13:21 crc kubenswrapper[4813]: I0317 09:13:21.730036 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:21 crc kubenswrapper[4813]: I0317 09:13:21.730039 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:21 crc kubenswrapper[4813]: E0317 09:13:21.730274 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:13:21 crc kubenswrapper[4813]: E0317 09:13:21.730345 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:13:21 crc kubenswrapper[4813]: I0317 09:13:21.730084 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:13:21 crc kubenswrapper[4813]: E0317 09:13:21.730477 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:13:22 crc kubenswrapper[4813]: I0317 09:13:22.730076 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:13:22 crc kubenswrapper[4813]: E0317 09:13:22.730334 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:13:22 crc kubenswrapper[4813]: I0317 09:13:22.731805 4813 scope.go:117] "RemoveContainer" containerID="09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b" Mar 17 09:13:23 crc kubenswrapper[4813]: I0317 09:13:23.607244 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-l47ql"] Mar 17 09:13:23 crc kubenswrapper[4813]: I0317 09:13:23.607799 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:13:23 crc kubenswrapper[4813]: E0317 09:13:23.607970 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:13:23 crc kubenswrapper[4813]: I0317 09:13:23.730936 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:23 crc kubenswrapper[4813]: I0317 09:13:23.731050 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:23 crc kubenswrapper[4813]: E0317 09:13:23.731112 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:13:23 crc kubenswrapper[4813]: E0317 09:13:23.731266 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:13:23 crc kubenswrapper[4813]: I0317 09:13:23.824526 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lngs6_a1730ca2-a1bb-4e9a-ba56-7edaded79f36/ovnkube-controller/3.log" Mar 17 09:13:23 crc kubenswrapper[4813]: I0317 09:13:23.827403 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerStarted","Data":"8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b"} Mar 17 09:13:23 crc kubenswrapper[4813]: I0317 09:13:23.828014 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:13:23 crc kubenswrapper[4813]: I0317 09:13:23.860711 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" podStartSLOduration=161.860686721 podStartE2EDuration="2m41.860686721s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:23.857346468 +0000 UTC m=+225.958149987" watchObservedRunningTime="2026-03-17 09:13:23.860686721 +0000 UTC m=+225.961490250" Mar 17 09:13:23 crc kubenswrapper[4813]: E0317 09:13:23.910740 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:13:24 crc kubenswrapper[4813]: I0317 09:13:24.730723 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:13:24 crc kubenswrapper[4813]: E0317 09:13:24.730904 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:13:25 crc kubenswrapper[4813]: I0317 09:13:25.729804 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:25 crc kubenswrapper[4813]: I0317 09:13:25.729847 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:25 crc kubenswrapper[4813]: E0317 09:13:25.730044 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:13:25 crc kubenswrapper[4813]: I0317 09:13:25.730076 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:13:25 crc kubenswrapper[4813]: E0317 09:13:25.730197 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:13:25 crc kubenswrapper[4813]: E0317 09:13:25.730346 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:13:26 crc kubenswrapper[4813]: I0317 09:13:26.730382 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:13:26 crc kubenswrapper[4813]: E0317 09:13:26.730579 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:13:27 crc kubenswrapper[4813]: I0317 09:13:27.730324 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:13:27 crc kubenswrapper[4813]: I0317 09:13:27.730339 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:27 crc kubenswrapper[4813]: E0317 09:13:27.730939 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:13:27 crc kubenswrapper[4813]: E0317 09:13:27.731240 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:13:27 crc kubenswrapper[4813]: I0317 09:13:27.730400 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:27 crc kubenswrapper[4813]: E0317 09:13:27.731420 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:13:28 crc kubenswrapper[4813]: I0317 09:13:28.729932 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:13:28 crc kubenswrapper[4813]: E0317 09:13:28.731695 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:13:28 crc kubenswrapper[4813]: E0317 09:13:28.911498 4813 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:13:29 crc kubenswrapper[4813]: I0317 09:13:29.571118 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:13:29 crc kubenswrapper[4813]: I0317 09:13:29.730577 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:13:29 crc kubenswrapper[4813]: I0317 09:13:29.730630 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:29 crc kubenswrapper[4813]: E0317 09:13:29.730730 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:13:29 crc kubenswrapper[4813]: I0317 09:13:29.730796 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:29 crc kubenswrapper[4813]: E0317 09:13:29.730953 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:13:29 crc kubenswrapper[4813]: E0317 09:13:29.731011 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:13:30 crc kubenswrapper[4813]: I0317 09:13:30.730090 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:13:30 crc kubenswrapper[4813]: E0317 09:13:30.730277 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:13:31 crc kubenswrapper[4813]: I0317 09:13:31.730350 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:13:31 crc kubenswrapper[4813]: I0317 09:13:31.730413 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:31 crc kubenswrapper[4813]: I0317 09:13:31.730364 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:31 crc kubenswrapper[4813]: E0317 09:13:31.730578 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:13:31 crc kubenswrapper[4813]: E0317 09:13:31.730735 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:13:31 crc kubenswrapper[4813]: E0317 09:13:31.730904 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:13:31 crc kubenswrapper[4813]: I0317 09:13:31.731373 4813 scope.go:117] "RemoveContainer" containerID="3cf6e87ac86b646f6e7e520f8dccf5788696ddef35970778f9490f817a04385a" Mar 17 09:13:32 crc kubenswrapper[4813]: I0317 09:13:32.729802 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:13:32 crc kubenswrapper[4813]: E0317 09:13:32.729973 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 17 09:13:32 crc kubenswrapper[4813]: I0317 09:13:32.862695 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5jjhb_c1246d4d-93d5-4a97-bef8-1ed881e1a217/kube-multus/1.log" Mar 17 09:13:32 crc kubenswrapper[4813]: I0317 09:13:32.863799 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5jjhb" event={"ID":"c1246d4d-93d5-4a97-bef8-1ed881e1a217","Type":"ContainerStarted","Data":"b72d7d564b1340f9238d2a5b0d3bda56b2ecbeb27116e4100944814dcef2fb8e"} Mar 17 09:13:33 crc kubenswrapper[4813]: I0317 09:13:33.730169 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:33 crc kubenswrapper[4813]: I0317 09:13:33.730300 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:13:33 crc kubenswrapper[4813]: I0317 09:13:33.730173 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:33 crc kubenswrapper[4813]: E0317 09:13:33.730422 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 17 09:13:33 crc kubenswrapper[4813]: E0317 09:13:33.730537 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-l47ql" podUID="27d4c184-5dd3-492a-b927-f7a7f31291ef" Mar 17 09:13:33 crc kubenswrapper[4813]: E0317 09:13:33.730785 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 17 09:13:34 crc kubenswrapper[4813]: I0317 09:13:34.729722 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:13:34 crc kubenswrapper[4813]: I0317 09:13:34.734017 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 17 09:13:34 crc kubenswrapper[4813]: I0317 09:13:34.734586 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 17 09:13:35 crc kubenswrapper[4813]: I0317 09:13:35.730129 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:35 crc kubenswrapper[4813]: I0317 09:13:35.730197 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:13:35 crc kubenswrapper[4813]: I0317 09:13:35.730527 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:35 crc kubenswrapper[4813]: I0317 09:13:35.733548 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 17 09:13:35 crc kubenswrapper[4813]: I0317 09:13:35.733553 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 17 09:13:35 crc kubenswrapper[4813]: I0317 09:13:35.733689 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 17 09:13:35 crc kubenswrapper[4813]: I0317 09:13:35.733758 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.317710 4813 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.360782 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kns96"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.361898 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.363722 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ptq22"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.364142 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.370548 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.370760 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.370794 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.370957 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-sk47b"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.371425 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-sk47b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.371798 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.372223 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.372672 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.373523 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.375264 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.376330 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.384654 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.384944 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.385192 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.386443 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.387030 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.387084 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.387204 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.387308 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.387364 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.387468 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.387489 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.387721 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.389967 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/26a86c9b-1645-49c2-a2fb-78b210984363-node-pullsecrets\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.390006 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.390061 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-knxnx\" (UniqueName: \"kubernetes.io/projected/feae3098-e93c-4dff-9088-a38db6a5d929-kube-api-access-knxnx\") pod \"machine-api-operator-5694c8668f-sk47b\" (UID: \"feae3098-e93c-4dff-9088-a38db6a5d929\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sk47b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.390123 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzxc7\" (UniqueName: \"kubernetes.io/projected/26a86c9b-1645-49c2-a2fb-78b210984363-kube-api-access-kzxc7\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.390165 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/26a86c9b-1645-49c2-a2fb-78b210984363-encryption-config\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.390234 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/feae3098-e93c-4dff-9088-a38db6a5d929-images\") pod \"machine-api-operator-5694c8668f-sk47b\" (UID: \"feae3098-e93c-4dff-9088-a38db6a5d929\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sk47b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.390296 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/feae3098-e93c-4dff-9088-a38db6a5d929-config\") pod \"machine-api-operator-5694c8668f-sk47b\" (UID: \"feae3098-e93c-4dff-9088-a38db6a5d929\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sk47b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.390670 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-tzt6n"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.390802 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26a86c9b-1645-49c2-a2fb-78b210984363-config\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.390915 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/26a86c9b-1645-49c2-a2fb-78b210984363-audit\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.391015 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/feae3098-e93c-4dff-9088-a38db6a5d929-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-sk47b\" (UID: \"feae3098-e93c-4dff-9088-a38db6a5d929\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sk47b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.391092 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/26a86c9b-1645-49c2-a2fb-78b210984363-audit-dir\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.391190 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/26a86c9b-1645-49c2-a2fb-78b210984363-etcd-serving-ca\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.391262 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/26a86c9b-1645-49c2-a2fb-78b210984363-image-import-ca\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.391331 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/26a86c9b-1645-49c2-a2fb-78b210984363-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.391397 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26a86c9b-1645-49c2-a2fb-78b210984363-serving-cert\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.391463 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/26a86c9b-1645-49c2-a2fb-78b210984363-etcd-client\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.391531 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-tzt6n" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.393367 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k58b"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.394354 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k58b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.395248 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.395588 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.396012 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.398799 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.399267 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.399349 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5kp7w"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.399413 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.399737 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.399785 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.399964 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.400284 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-kpxlg"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.400939 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kpxlg" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.400967 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.403232 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.403335 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.407521 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.411888 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-4rqzj"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.412654 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4rqzj" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.412709 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.413372 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dpghh"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.415090 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.428284 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.429005 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wcd9z"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.430495 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dpghh" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.430725 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.431017 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.432819 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.432867 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.436006 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.446807 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.446884 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.446997 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.447469 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.447558 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.447658 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.447842 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.447991 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.448058 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.448118 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.448192 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.448225 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.448337 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.448863 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2fmh2"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.448964 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.449052 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.449140 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.449214 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.449311 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-2fmh2" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.449360 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.449614 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wcd9z" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.449869 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-xb4k9"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.450280 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.450677 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-xb4k9" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.450758 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.450871 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.450969 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.451059 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.451200 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.451222 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.451344 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.451378 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.451504 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.452241 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.454347 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.454378 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-qwlbl"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.454837 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-njr98"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.455082 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-m97th"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.455369 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-m97th" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.455695 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-qwlbl" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.455867 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-njr98" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.456720 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.456817 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s2hhg"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.461994 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.462435 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.464427 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.464792 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s2hhg" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.466435 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-jzbwp"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.467240 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.467664 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-2w9vb"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.468332 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.471737 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dbm4b"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.472434 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.475299 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.480197 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwbwr"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.480728 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwbwr" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.484924 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.485089 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.485097 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.485218 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.485310 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.485398 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.485541 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.485719 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.485826 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.485924 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.486016 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.486089 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.486102 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.486176 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.486180 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.486179 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.486244 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.486282 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.486337 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.486350 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.486414 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.486430 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.486475 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.486546 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.486584 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.486672 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.486741 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.486764 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.487117 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.500946 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.502647 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-tbl5t"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.512091 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-config\") pod \"controller-manager-879f6c89f-ptq22\" (UID: \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.512138 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26a86c9b-1645-49c2-a2fb-78b210984363-serving-cert\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.512156 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h42nb\" (UniqueName: \"kubernetes.io/projected/eb147d37-445a-48f2-a2e6-fb16f3575529-kube-api-access-h42nb\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.512174 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-serving-cert\") pod \"controller-manager-879f6c89f-ptq22\" (UID: \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.512190 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.512207 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eb147d37-445a-48f2-a2e6-fb16f3575529-audit-dir\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.512223 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/768d1a31-0d78-4239-9d7d-260d11e4ad58-client-ca\") pod \"route-controller-manager-6576b87f9c-6rn9l\" (UID: \"768d1a31-0d78-4239-9d7d-260d11e4ad58\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.512241 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/26a86c9b-1645-49c2-a2fb-78b210984363-etcd-client\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.512256 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eb147d37-445a-48f2-a2e6-fb16f3575529-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.512271 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/768d1a31-0d78-4239-9d7d-260d11e4ad58-serving-cert\") pod \"route-controller-manager-6576b87f9c-6rn9l\" (UID: \"768d1a31-0d78-4239-9d7d-260d11e4ad58\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.512949 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6m97\" (UniqueName: \"kubernetes.io/projected/768d1a31-0d78-4239-9d7d-260d11e4ad58-kube-api-access-d6m97\") pod \"route-controller-manager-6576b87f9c-6rn9l\" (UID: \"768d1a31-0d78-4239-9d7d-260d11e4ad58\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.513000 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5djq"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.513115 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tbl5t" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.513530 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5djq" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.513767 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/26a86c9b-1645-49c2-a2fb-78b210984363-node-pullsecrets\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.513826 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/26a86c9b-1645-49c2-a2fb-78b210984363-node-pullsecrets\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.513849 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eb147d37-445a-48f2-a2e6-fb16f3575529-audit-policies\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.513869 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-knxnx\" (UniqueName: \"kubernetes.io/projected/feae3098-e93c-4dff-9088-a38db6a5d929-kube-api-access-knxnx\") pod \"machine-api-operator-5694c8668f-sk47b\" (UID: \"feae3098-e93c-4dff-9088-a38db6a5d929\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sk47b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.513905 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98211622-4794-4412-a9d0-64436ee1a3f4-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2k58b\" (UID: \"98211622-4794-4412-a9d0-64436ee1a3f4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k58b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.513919 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-audit-dir\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.513934 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.513950 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/768d1a31-0d78-4239-9d7d-260d11e4ad58-config\") pod \"route-controller-manager-6576b87f9c-6rn9l\" (UID: \"768d1a31-0d78-4239-9d7d-260d11e4ad58\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.513967 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzxc7\" (UniqueName: \"kubernetes.io/projected/26a86c9b-1645-49c2-a2fb-78b210984363-kube-api-access-kzxc7\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.513981 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct9wj\" (UniqueName: \"kubernetes.io/projected/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-kube-api-access-ct9wj\") pod \"controller-manager-879f6c89f-ptq22\" (UID: \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.513996 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/26a86c9b-1645-49c2-a2fb-78b210984363-encryption-config\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.514011 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.514043 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.514058 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.514076 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/eb147d37-445a-48f2-a2e6-fb16f3575529-encryption-config\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.514093 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/feae3098-e93c-4dff-9088-a38db6a5d929-images\") pod \"machine-api-operator-5694c8668f-sk47b\" (UID: \"feae3098-e93c-4dff-9088-a38db6a5d929\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sk47b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.514110 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/eb147d37-445a-48f2-a2e6-fb16f3575529-etcd-client\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.514125 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-client-ca\") pod \"controller-manager-879f6c89f-ptq22\" (UID: \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.514142 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-audit-policies\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.514163 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/feae3098-e93c-4dff-9088-a38db6a5d929-config\") pod \"machine-api-operator-5694c8668f-sk47b\" (UID: \"feae3098-e93c-4dff-9088-a38db6a5d929\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sk47b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.514178 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ptq22\" (UID: \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.514195 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26a86c9b-1645-49c2-a2fb-78b210984363-config\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.514209 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/26a86c9b-1645-49c2-a2fb-78b210984363-audit\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.514224 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/eb147d37-445a-48f2-a2e6-fb16f3575529-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.514241 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/feae3098-e93c-4dff-9088-a38db6a5d929-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-sk47b\" (UID: \"feae3098-e93c-4dff-9088-a38db6a5d929\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sk47b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.514260 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98211622-4794-4412-a9d0-64436ee1a3f4-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2k58b\" (UID: \"98211622-4794-4412-a9d0-64436ee1a3f4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k58b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.514274 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb147d37-445a-48f2-a2e6-fb16f3575529-serving-cert\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.514291 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/26a86c9b-1645-49c2-a2fb-78b210984363-audit-dir\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.514323 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/26a86c9b-1645-49c2-a2fb-78b210984363-etcd-serving-ca\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.514337 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/26a86c9b-1645-49c2-a2fb-78b210984363-image-import-ca\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.514352 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/26a86c9b-1645-49c2-a2fb-78b210984363-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.514368 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kntnz\" (UniqueName: \"kubernetes.io/projected/98211622-4794-4412-a9d0-64436ee1a3f4-kube-api-access-kntnz\") pod \"openshift-apiserver-operator-796bbdcf4f-2k58b\" (UID: \"98211622-4794-4412-a9d0-64436ee1a3f4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k58b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.515175 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lrhrs"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.515828 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lrhrs" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.515878 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/feae3098-e93c-4dff-9088-a38db6a5d929-images\") pod \"machine-api-operator-5694c8668f-sk47b\" (UID: \"feae3098-e93c-4dff-9088-a38db6a5d929\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sk47b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.516687 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/26a86c9b-1645-49c2-a2fb-78b210984363-config\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.516771 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/26a86c9b-1645-49c2-a2fb-78b210984363-image-import-ca\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.516816 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/26a86c9b-1645-49c2-a2fb-78b210984363-audit-dir\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.516906 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vm474"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.517525 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vm474" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.518206 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/feae3098-e93c-4dff-9088-a38db6a5d929-config\") pod \"machine-api-operator-5694c8668f-sk47b\" (UID: \"feae3098-e93c-4dff-9088-a38db6a5d929\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sk47b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.518233 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/26a86c9b-1645-49c2-a2fb-78b210984363-trusted-ca-bundle\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.518417 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/26a86c9b-1645-49c2-a2fb-78b210984363-audit\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.518485 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.518807 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/26a86c9b-1645-49c2-a2fb-78b210984363-etcd-serving-ca\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.523122 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/26a86c9b-1645-49c2-a2fb-78b210984363-encryption-config\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.524579 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.526070 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/26a86c9b-1645-49c2-a2fb-78b210984363-etcd-client\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.529654 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5wl9g"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.530170 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6lsz5"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.530587 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-6lsz5" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.530826 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.531045 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bpfb"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.531334 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bpfb" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.532393 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8qg9n"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.532999 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8qg9n" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.533216 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/26a86c9b-1645-49c2-a2fb-78b210984363-serving-cert\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.533300 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562312-hjpcp"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.534213 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562312-hjpcp" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.534399 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-s6zwp"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.534966 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s6zwp" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.535652 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wttgv"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.536119 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-wttgv" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.537150 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/feae3098-e93c-4dff-9088-a38db6a5d929-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-sk47b\" (UID: \"feae3098-e93c-4dff-9088-a38db6a5d929\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sk47b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.540015 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-ghwhk"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.542218 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7w9l5"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.542961 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ghwhk" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.543170 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-7tdh4"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.542973 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7w9l5" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.543918 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7tdh4" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.545036 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.545284 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.545821 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.546652 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tjvgc"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.546922 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.547309 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.549023 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kns96"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.549166 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tjvgc" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.549051 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-sk47b"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.560247 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ptq22"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.561203 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-tzt6n"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.562120 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.563436 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k58b"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.564517 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.571904 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-8hxbx"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.572841 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8hxbx" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.576980 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2fmh2"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.577929 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5kp7w"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.579179 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5djq"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.580209 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wcd9z"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.581276 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-qwlbl"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.582954 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwbwr"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.583480 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dbm4b"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.584490 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.584469 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dpghh"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.585692 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-2w9vb"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.587003 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.588652 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bpfb"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.590512 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8qg9n"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.592302 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-njr98"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.593785 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-4rqzj"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.597796 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-jzbwp"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.601656 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lrhrs"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.604547 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.605721 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wttgv"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.617450 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-tbl5t"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.618468 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kntnz\" (UniqueName: \"kubernetes.io/projected/98211622-4794-4412-a9d0-64436ee1a3f4-kube-api-access-kntnz\") pod \"openshift-apiserver-operator-796bbdcf4f-2k58b\" (UID: \"98211622-4794-4412-a9d0-64436ee1a3f4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k58b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.618581 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-config\") pod \"controller-manager-879f6c89f-ptq22\" (UID: \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.618727 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h42nb\" (UniqueName: \"kubernetes.io/projected/eb147d37-445a-48f2-a2e6-fb16f3575529-kube-api-access-h42nb\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.618835 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.618976 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.619087 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-serving-cert\") pod \"controller-manager-879f6c89f-ptq22\" (UID: \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.619196 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.619476 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eb147d37-445a-48f2-a2e6-fb16f3575529-audit-dir\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.619630 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/768d1a31-0d78-4239-9d7d-260d11e4ad58-client-ca\") pod \"route-controller-manager-6576b87f9c-6rn9l\" (UID: \"768d1a31-0d78-4239-9d7d-260d11e4ad58\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.619742 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eb147d37-445a-48f2-a2e6-fb16f3575529-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.619836 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/768d1a31-0d78-4239-9d7d-260d11e4ad58-serving-cert\") pod \"route-controller-manager-6576b87f9c-6rn9l\" (UID: \"768d1a31-0d78-4239-9d7d-260d11e4ad58\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.620375 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.620459 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eb147d37-445a-48f2-a2e6-fb16f3575529-audit-policies\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.620537 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6m97\" (UniqueName: \"kubernetes.io/projected/768d1a31-0d78-4239-9d7d-260d11e4ad58-kube-api-access-d6m97\") pod \"route-controller-manager-6576b87f9c-6rn9l\" (UID: \"768d1a31-0d78-4239-9d7d-260d11e4ad58\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.620656 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98211622-4794-4412-a9d0-64436ee1a3f4-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2k58b\" (UID: \"98211622-4794-4412-a9d0-64436ee1a3f4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k58b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.620739 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-audit-dir\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.620818 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.620898 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/768d1a31-0d78-4239-9d7d-260d11e4ad58-config\") pod \"route-controller-manager-6576b87f9c-6rn9l\" (UID: \"768d1a31-0d78-4239-9d7d-260d11e4ad58\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.620982 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct9wj\" (UniqueName: \"kubernetes.io/projected/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-kube-api-access-ct9wj\") pod \"controller-manager-879f6c89f-ptq22\" (UID: \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621047 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-audit-dir\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621059 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/09bf1deb-b09d-4978-baac-c06782c2ca2f-trusted-ca\") pod \"console-operator-58897d9998-2fmh2\" (UID: \"09bf1deb-b09d-4978-baac-c06782c2ca2f\") " pod="openshift-console-operator/console-operator-58897d9998-2fmh2" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621140 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621205 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621232 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621251 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/eb147d37-445a-48f2-a2e6-fb16f3575529-encryption-config\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621294 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cw28k\" (UniqueName: \"kubernetes.io/projected/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-kube-api-access-cw28k\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621323 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-client-ca\") pod \"controller-manager-879f6c89f-ptq22\" (UID: \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621341 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09bf1deb-b09d-4978-baac-c06782c2ca2f-serving-cert\") pod \"console-operator-58897d9998-2fmh2\" (UID: \"09bf1deb-b09d-4978-baac-c06782c2ca2f\") " pod="openshift-console-operator/console-operator-58897d9998-2fmh2" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621371 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/eb147d37-445a-48f2-a2e6-fb16f3575529-etcd-client\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621401 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-audit-policies\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621429 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621458 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621496 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ptq22\" (UID: \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621523 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621584 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/eb147d37-445a-48f2-a2e6-fb16f3575529-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621645 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09bf1deb-b09d-4978-baac-c06782c2ca2f-config\") pod \"console-operator-58897d9998-2fmh2\" (UID: \"09bf1deb-b09d-4978-baac-c06782c2ca2f\") " pod="openshift-console-operator/console-operator-58897d9998-2fmh2" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621674 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z55br\" (UniqueName: \"kubernetes.io/projected/09bf1deb-b09d-4978-baac-c06782c2ca2f-kube-api-access-z55br\") pod \"console-operator-58897d9998-2fmh2\" (UID: \"09bf1deb-b09d-4978-baac-c06782c2ca2f\") " pod="openshift-console-operator/console-operator-58897d9998-2fmh2" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621704 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98211622-4794-4412-a9d0-64436ee1a3f4-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2k58b\" (UID: \"98211622-4794-4412-a9d0-64436ee1a3f4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k58b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621738 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb147d37-445a-48f2-a2e6-fb16f3575529-serving-cert\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621754 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/98211622-4794-4412-a9d0-64436ee1a3f4-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2k58b\" (UID: \"98211622-4794-4412-a9d0-64436ee1a3f4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k58b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.623534 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/eb147d37-445a-48f2-a2e6-fb16f3575529-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.622306 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/eb147d37-445a-48f2-a2e6-fb16f3575529-audit-policies\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.620492 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/eb147d37-445a-48f2-a2e6-fb16f3575529-audit-dir\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.622346 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.622641 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-serving-cert\") pod \"controller-manager-879f6c89f-ptq22\" (UID: \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.623193 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/768d1a31-0d78-4239-9d7d-260d11e4ad58-config\") pod \"route-controller-manager-6576b87f9c-6rn9l\" (UID: \"768d1a31-0d78-4239-9d7d-260d11e4ad58\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.620734 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-config\") pod \"controller-manager-879f6c89f-ptq22\" (UID: \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.621948 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/768d1a31-0d78-4239-9d7d-260d11e4ad58-client-ca\") pod \"route-controller-manager-6576b87f9c-6rn9l\" (UID: \"768d1a31-0d78-4239-9d7d-260d11e4ad58\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.618887 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-xb4k9"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.624325 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-7tdh4"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.624344 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-cjn8k"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.624424 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.626270 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-cjn8k" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.626565 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-gczkp"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.636874 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6lsz5"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.637350 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-gczkp" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.637968 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/eb147d37-445a-48f2-a2e6-fb16f3575529-etcd-client\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.638147 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.638372 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/98211622-4794-4412-a9d0-64436ee1a3f4-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2k58b\" (UID: \"98211622-4794-4412-a9d0-64436ee1a3f4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k58b" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.638442 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.638780 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/eb147d37-445a-48f2-a2e6-fb16f3575529-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.639045 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-ptq22\" (UID: \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.639446 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/768d1a31-0d78-4239-9d7d-260d11e4ad58-serving-cert\") pod \"route-controller-manager-6576b87f9c-6rn9l\" (UID: \"768d1a31-0d78-4239-9d7d-260d11e4ad58\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.639453 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-client-ca\") pod \"controller-manager-879f6c89f-ptq22\" (UID: \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.639689 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.640095 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-audit-policies\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.641534 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/eb147d37-445a-48f2-a2e6-fb16f3575529-encryption-config\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.641609 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.641840 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/eb147d37-445a-48f2-a2e6-fb16f3575529-serving-cert\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.642285 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.643613 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.643709 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-ghwhk"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.645319 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5wl9g"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.646826 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-s6zwp"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.648135 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-cjn8k"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.649159 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s2hhg"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.650446 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7w9l5"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.651764 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562312-hjpcp"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.652774 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vm474"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.654379 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tjvgc"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.655942 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-gczkp"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.657630 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.658803 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-rw942"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.660528 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-rw942"] Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.660636 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-rw942" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.664855 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.683809 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.703655 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.723818 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.724108 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/09bf1deb-b09d-4978-baac-c06782c2ca2f-trusted-ca\") pod \"console-operator-58897d9998-2fmh2\" (UID: \"09bf1deb-b09d-4978-baac-c06782c2ca2f\") " pod="openshift-console-operator/console-operator-58897d9998-2fmh2" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.724149 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cw28k\" (UniqueName: \"kubernetes.io/projected/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-kube-api-access-cw28k\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.724167 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09bf1deb-b09d-4978-baac-c06782c2ca2f-serving-cert\") pod \"console-operator-58897d9998-2fmh2\" (UID: \"09bf1deb-b09d-4978-baac-c06782c2ca2f\") " pod="openshift-console-operator/console-operator-58897d9998-2fmh2" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.724185 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.724207 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.724307 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.724332 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09bf1deb-b09d-4978-baac-c06782c2ca2f-config\") pod \"console-operator-58897d9998-2fmh2\" (UID: \"09bf1deb-b09d-4978-baac-c06782c2ca2f\") " pod="openshift-console-operator/console-operator-58897d9998-2fmh2" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.724348 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z55br\" (UniqueName: \"kubernetes.io/projected/09bf1deb-b09d-4978-baac-c06782c2ca2f-kube-api-access-z55br\") pod \"console-operator-58897d9998-2fmh2\" (UID: \"09bf1deb-b09d-4978-baac-c06782c2ca2f\") " pod="openshift-console-operator/console-operator-58897d9998-2fmh2" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.724393 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.724406 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.724424 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.725382 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.726078 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09bf1deb-b09d-4978-baac-c06782c2ca2f-config\") pod \"console-operator-58897d9998-2fmh2\" (UID: \"09bf1deb-b09d-4978-baac-c06782c2ca2f\") " pod="openshift-console-operator/console-operator-58897d9998-2fmh2" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.726697 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/09bf1deb-b09d-4978-baac-c06782c2ca2f-trusted-ca\") pod \"console-operator-58897d9998-2fmh2\" (UID: \"09bf1deb-b09d-4978-baac-c06782c2ca2f\") " pod="openshift-console-operator/console-operator-58897d9998-2fmh2" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.728199 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09bf1deb-b09d-4978-baac-c06782c2ca2f-serving-cert\") pod \"console-operator-58897d9998-2fmh2\" (UID: \"09bf1deb-b09d-4978-baac-c06782c2ca2f\") " pod="openshift-console-operator/console-operator-58897d9998-2fmh2" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.728202 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.728318 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.728765 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.729694 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.729875 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.743717 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.764426 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.784522 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.803954 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.823869 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.851465 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.863943 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.884538 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.904631 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.924383 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.943891 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.963411 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 17 09:13:41 crc kubenswrapper[4813]: I0317 09:13:41.983646 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.003836 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.025303 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.044665 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.064709 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.084289 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.103811 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.164588 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.185160 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.204519 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.225021 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.244975 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.292321 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-knxnx\" (UniqueName: \"kubernetes.io/projected/feae3098-e93c-4dff-9088-a38db6a5d929-kube-api-access-knxnx\") pod \"machine-api-operator-5694c8668f-sk47b\" (UID: \"feae3098-e93c-4dff-9088-a38db6a5d929\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sk47b" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.304319 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.305261 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzxc7\" (UniqueName: \"kubernetes.io/projected/26a86c9b-1645-49c2-a2fb-78b210984363-kube-api-access-kzxc7\") pod \"apiserver-76f77b778f-kns96\" (UID: \"26a86c9b-1645-49c2-a2fb-78b210984363\") " pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.324772 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.344999 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.357787 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-sk47b" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.365738 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.423495 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.423841 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.424155 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.444282 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.464299 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.484450 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.505526 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.536797 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.542980 4813 request.go:700] Waited for 1.011960042s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/secrets?fieldSelector=metadata.name%3Dmarketplace-operator-metrics&limit=500&resourceVersion=0 Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.544935 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.574634 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.585761 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.601496 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.604208 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.625249 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.644647 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.664358 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.685222 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.701737 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-sk47b"] Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.704116 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.724472 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.745520 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.766152 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.786067 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.804635 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.825764 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.843507 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.864316 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.869436 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-kns96"] Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.894267 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.894808 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-sk47b" event={"ID":"feae3098-e93c-4dff-9088-a38db6a5d929","Type":"ContainerStarted","Data":"df0eb00410343eeeb7c1728a5c7ac5226aaf84b3ff791c3b97dcf771d28e962a"} Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.894860 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-sk47b" event={"ID":"feae3098-e93c-4dff-9088-a38db6a5d929","Type":"ContainerStarted","Data":"e635fb9abf1236b7ab3a381b179742d7e7c242544989849f6391114534e46f0d"} Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.895797 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kns96" event={"ID":"26a86c9b-1645-49c2-a2fb-78b210984363","Type":"ContainerStarted","Data":"d04fca3ff535644a7872e52f3e658a6658e24f3fe43bc1254b70f480173880d9"} Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.904268 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.924827 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.943868 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.964562 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 17 09:13:42 crc kubenswrapper[4813]: I0317 09:13:42.984775 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.004416 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.024495 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.043950 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.064268 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.084390 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.104744 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.124201 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.143912 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.164453 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.184227 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.204564 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.224676 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.244633 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.264675 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.284432 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.304520 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.354872 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kntnz\" (UniqueName: \"kubernetes.io/projected/98211622-4794-4412-a9d0-64436ee1a3f4-kube-api-access-kntnz\") pod \"openshift-apiserver-operator-796bbdcf4f-2k58b\" (UID: \"98211622-4794-4412-a9d0-64436ee1a3f4\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k58b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.364519 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h42nb\" (UniqueName: \"kubernetes.io/projected/eb147d37-445a-48f2-a2e6-fb16f3575529-kube-api-access-h42nb\") pod \"apiserver-7bbb656c7d-h87sm\" (UID: \"eb147d37-445a-48f2-a2e6-fb16f3575529\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.393436 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6m97\" (UniqueName: \"kubernetes.io/projected/768d1a31-0d78-4239-9d7d-260d11e4ad58-kube-api-access-d6m97\") pod \"route-controller-manager-6576b87f9c-6rn9l\" (UID: \"768d1a31-0d78-4239-9d7d-260d11e4ad58\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.403103 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct9wj\" (UniqueName: \"kubernetes.io/projected/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-kube-api-access-ct9wj\") pod \"controller-manager-879f6c89f-ptq22\" (UID: \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\") " pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.404377 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.424935 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.444843 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.467999 4813 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.485686 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.507086 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.524434 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.545408 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.548425 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.564916 4813 request.go:700] Waited for 1.90399402s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress-canary/configmaps?fieldSelector=metadata.name%3Dkube-root-ca.crt&limit=500&resourceVersion=0 Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.571248 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.577318 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.584439 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.610447 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.620043 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z55br\" (UniqueName: \"kubernetes.io/projected/09bf1deb-b09d-4978-baac-c06782c2ca2f-kube-api-access-z55br\") pod \"console-operator-58897d9998-2fmh2\" (UID: \"09bf1deb-b09d-4978-baac-c06782c2ca2f\") " pod="openshift-console-operator/console-operator-58897d9998-2fmh2" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.637683 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cw28k\" (UniqueName: \"kubernetes.io/projected/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-kube-api-access-cw28k\") pod \"oauth-openshift-558db77b4-5kp7w\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.653037 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k58b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.662412 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.704291 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-2fmh2" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.753569 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/782b4936-2caa-44da-8716-c57924be6df3-registry-certificates\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.753645 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/93cb7b2d-1c40-40be-82d7-1858272444ec-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-njr98\" (UID: \"93cb7b2d-1c40-40be-82d7-1858272444ec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-njr98" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.753674 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdnwc\" (UniqueName: \"kubernetes.io/projected/782b4936-2caa-44da-8716-c57924be6df3-kube-api-access-kdnwc\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.753696 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/68399978-6f43-4458-a1e1-58d5984de8f5-service-ca-bundle\") pod \"router-default-5444994796-m97th\" (UID: \"68399978-6f43-4458-a1e1-58d5984de8f5\") " pod="openshift-ingress/router-default-5444994796-m97th" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.753727 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/782b4936-2caa-44da-8716-c57924be6df3-ca-trust-extracted\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.753748 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f7603094-bb1d-4d77-b463-2270003b2805-console-serving-cert\") pod \"console-f9d7485db-2w9vb\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.753787 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24qg4\" (UniqueName: \"kubernetes.io/projected/3aa0fc32-ed11-482b-9a6d-fcd59544bf8d-kube-api-access-24qg4\") pod \"downloads-7954f5f757-xb4k9\" (UID: \"3aa0fc32-ed11-482b-9a6d-fcd59544bf8d\") " pod="openshift-console/downloads-7954f5f757-xb4k9" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.753807 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/364cd6ac-033f-455e-a95a-10d50f055fb2-etcd-client\") pod \"etcd-operator-b45778765-jzbwp\" (UID: \"364cd6ac-033f-455e-a95a-10d50f055fb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.753840 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4lhgk\" (UniqueName: \"kubernetes.io/projected/68399978-6f43-4458-a1e1-58d5984de8f5-kube-api-access-4lhgk\") pod \"router-default-5444994796-m97th\" (UID: \"68399978-6f43-4458-a1e1-58d5984de8f5\") " pod="openshift-ingress/router-default-5444994796-m97th" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.753886 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cnpcn\" (UniqueName: \"kubernetes.io/projected/845ee13b-7ed9-4910-958b-5269812237f0-kube-api-access-cnpcn\") pod \"cluster-samples-operator-665b6dd947-dpghh\" (UID: \"845ee13b-7ed9-4910-958b-5269812237f0\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dpghh" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.753908 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12efeac2-9fc3-425a-8874-3c2d957e04ca-config\") pod \"machine-approver-56656f9798-kpxlg\" (UID: \"12efeac2-9fc3-425a-8874-3c2d957e04ca\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kpxlg" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754086 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/68399978-6f43-4458-a1e1-58d5984de8f5-stats-auth\") pod \"router-default-5444994796-m97th\" (UID: \"68399978-6f43-4458-a1e1-58d5984de8f5\") " pod="openshift-ingress/router-default-5444994796-m97th" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754116 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dx7gs\" (UniqueName: \"kubernetes.io/projected/364cd6ac-033f-455e-a95a-10d50f055fb2-kube-api-access-dx7gs\") pod \"etcd-operator-b45778765-jzbwp\" (UID: \"364cd6ac-033f-455e-a95a-10d50f055fb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754137 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68399978-6f43-4458-a1e1-58d5984de8f5-metrics-certs\") pod \"router-default-5444994796-m97th\" (UID: \"68399978-6f43-4458-a1e1-58d5984de8f5\") " pod="openshift-ingress/router-default-5444994796-m97th" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754163 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72fa1c68-872b-45a3-98b7-b0b1ebc08ad7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wcd9z\" (UID: \"72fa1c68-872b-45a3-98b7-b0b1ebc08ad7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wcd9z" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754207 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bd4438d-8a23-4e8b-aad7-f5775f1d88a8-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wwbwr\" (UID: \"2bd4438d-8a23-4e8b-aad7-f5775f1d88a8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwbwr" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754230 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/550c18c5-f4e0-4da1-9692-c6cacfa882af-config\") pod \"authentication-operator-69f744f599-tzt6n\" (UID: \"550c18c5-f4e0-4da1-9692-c6cacfa882af\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tzt6n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754251 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/364cd6ac-033f-455e-a95a-10d50f055fb2-etcd-service-ca\") pod \"etcd-operator-b45778765-jzbwp\" (UID: \"364cd6ac-033f-455e-a95a-10d50f055fb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754280 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/364cd6ac-033f-455e-a95a-10d50f055fb2-config\") pod \"etcd-operator-b45778765-jzbwp\" (UID: \"364cd6ac-033f-455e-a95a-10d50f055fb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754300 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8745x\" (UniqueName: \"kubernetes.io/projected/93cb7b2d-1c40-40be-82d7-1858272444ec-kube-api-access-8745x\") pod \"cluster-image-registry-operator-dc59b4c8b-njr98\" (UID: \"93cb7b2d-1c40-40be-82d7-1858272444ec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-njr98" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754331 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2bd4438d-8a23-4e8b-aad7-f5775f1d88a8-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wwbwr\" (UID: \"2bd4438d-8a23-4e8b-aad7-f5775f1d88a8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwbwr" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754351 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/782b4936-2caa-44da-8716-c57924be6df3-registry-tls\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754371 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b828a0e2-3def-4bb5-8f54-2792376423cd-serving-cert\") pod \"openshift-config-operator-7777fb866f-4rqzj\" (UID: \"b828a0e2-3def-4bb5-8f54-2792376423cd\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4rqzj" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754393 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/12efeac2-9fc3-425a-8874-3c2d957e04ca-auth-proxy-config\") pod \"machine-approver-56656f9798-kpxlg\" (UID: \"12efeac2-9fc3-425a-8874-3c2d957e04ca\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kpxlg" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754418 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2657\" (UniqueName: \"kubernetes.io/projected/550c18c5-f4e0-4da1-9692-c6cacfa882af-kube-api-access-z2657\") pod \"authentication-operator-69f744f599-tzt6n\" (UID: \"550c18c5-f4e0-4da1-9692-c6cacfa882af\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tzt6n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754442 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjfgf\" (UniqueName: \"kubernetes.io/projected/b828a0e2-3def-4bb5-8f54-2792376423cd-kube-api-access-xjfgf\") pod \"openshift-config-operator-7777fb866f-4rqzj\" (UID: \"b828a0e2-3def-4bb5-8f54-2792376423cd\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4rqzj" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754465 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-trusted-ca-bundle\") pod \"console-f9d7485db-2w9vb\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754498 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/550c18c5-f4e0-4da1-9692-c6cacfa882af-service-ca-bundle\") pod \"authentication-operator-69f744f599-tzt6n\" (UID: \"550c18c5-f4e0-4da1-9692-c6cacfa882af\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tzt6n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754517 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/94f282fa-7b22-41e3-9911-073b05d41200-metrics-tls\") pod \"dns-operator-744455d44c-qwlbl\" (UID: \"94f282fa-7b22-41e3-9911-073b05d41200\") " pod="openshift-dns-operator/dns-operator-744455d44c-qwlbl" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754535 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72fa1c68-872b-45a3-98b7-b0b1ebc08ad7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wcd9z\" (UID: \"72fa1c68-872b-45a3-98b7-b0b1ebc08ad7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wcd9z" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754557 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/782b4936-2caa-44da-8716-c57924be6df3-bound-sa-token\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754578 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b828a0e2-3def-4bb5-8f54-2792376423cd-available-featuregates\") pod \"openshift-config-operator-7777fb866f-4rqzj\" (UID: \"b828a0e2-3def-4bb5-8f54-2792376423cd\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4rqzj" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754613 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d32fb60f-2d76-4e02-9839-535e1543eff7-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-s2hhg\" (UID: \"d32fb60f-2d76-4e02-9839-535e1543eff7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s2hhg" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754637 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d32fb60f-2d76-4e02-9839-535e1543eff7-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-s2hhg\" (UID: \"d32fb60f-2d76-4e02-9839-535e1543eff7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s2hhg" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754680 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/93cb7b2d-1c40-40be-82d7-1858272444ec-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-njr98\" (UID: \"93cb7b2d-1c40-40be-82d7-1858272444ec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-njr98" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754704 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/782b4936-2caa-44da-8716-c57924be6df3-trusted-ca\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754722 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/550c18c5-f4e0-4da1-9692-c6cacfa882af-serving-cert\") pod \"authentication-operator-69f744f599-tzt6n\" (UID: \"550c18c5-f4e0-4da1-9692-c6cacfa882af\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tzt6n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754743 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s29cv\" (UniqueName: \"kubernetes.io/projected/12efeac2-9fc3-425a-8874-3c2d957e04ca-kube-api-access-s29cv\") pod \"machine-approver-56656f9798-kpxlg\" (UID: \"12efeac2-9fc3-425a-8874-3c2d957e04ca\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kpxlg" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754764 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/93cb7b2d-1c40-40be-82d7-1858272444ec-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-njr98\" (UID: \"93cb7b2d-1c40-40be-82d7-1858272444ec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-njr98" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754796 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/845ee13b-7ed9-4910-958b-5269812237f0-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-dpghh\" (UID: \"845ee13b-7ed9-4910-958b-5269812237f0\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dpghh" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754819 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/364cd6ac-033f-455e-a95a-10d50f055fb2-serving-cert\") pod \"etcd-operator-b45778765-jzbwp\" (UID: \"364cd6ac-033f-455e-a95a-10d50f055fb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754839 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/364cd6ac-033f-455e-a95a-10d50f055fb2-etcd-ca\") pod \"etcd-operator-b45778765-jzbwp\" (UID: \"364cd6ac-033f-455e-a95a-10d50f055fb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754872 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-console-config\") pod \"console-f9d7485db-2w9vb\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754912 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754935 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/550c18c5-f4e0-4da1-9692-c6cacfa882af-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-tzt6n\" (UID: \"550c18c5-f4e0-4da1-9692-c6cacfa882af\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tzt6n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754955 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/12efeac2-9fc3-425a-8874-3c2d957e04ca-machine-approver-tls\") pod \"machine-approver-56656f9798-kpxlg\" (UID: \"12efeac2-9fc3-425a-8874-3c2d957e04ca\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kpxlg" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.754976 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-service-ca\") pod \"console-f9d7485db-2w9vb\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.755012 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/68399978-6f43-4458-a1e1-58d5984de8f5-default-certificate\") pod \"router-default-5444994796-m97th\" (UID: \"68399978-6f43-4458-a1e1-58d5984de8f5\") " pod="openshift-ingress/router-default-5444994796-m97th" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.755031 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f7603094-bb1d-4d77-b463-2270003b2805-console-oauth-config\") pod \"console-f9d7485db-2w9vb\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.755052 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzrsb\" (UniqueName: \"kubernetes.io/projected/72fa1c68-872b-45a3-98b7-b0b1ebc08ad7-kube-api-access-tzrsb\") pod \"openshift-controller-manager-operator-756b6f6bc6-wcd9z\" (UID: \"72fa1c68-872b-45a3-98b7-b0b1ebc08ad7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wcd9z" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.755096 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngkkm\" (UniqueName: \"kubernetes.io/projected/f7603094-bb1d-4d77-b463-2270003b2805-kube-api-access-ngkkm\") pod \"console-f9d7485db-2w9vb\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.755121 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/782b4936-2caa-44da-8716-c57924be6df3-installation-pull-secrets\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.755189 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f2q6\" (UniqueName: \"kubernetes.io/projected/94f282fa-7b22-41e3-9911-073b05d41200-kube-api-access-7f2q6\") pod \"dns-operator-744455d44c-qwlbl\" (UID: \"94f282fa-7b22-41e3-9911-073b05d41200\") " pod="openshift-dns-operator/dns-operator-744455d44c-qwlbl" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.755239 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2bd4438d-8a23-4e8b-aad7-f5775f1d88a8-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wwbwr\" (UID: \"2bd4438d-8a23-4e8b-aad7-f5775f1d88a8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwbwr" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.755262 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d32fb60f-2d76-4e02-9839-535e1543eff7-config\") pod \"kube-apiserver-operator-766d6c64bb-s2hhg\" (UID: \"d32fb60f-2d76-4e02-9839-535e1543eff7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s2hhg" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.755286 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-oauth-serving-cert\") pod \"console-f9d7485db-2w9vb\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:43 crc kubenswrapper[4813]: E0317 09:13:43.756124 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:44.256108572 +0000 UTC m=+246.356912161 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.817795 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ptq22"] Mar 17 09:13:43 crc kubenswrapper[4813]: W0317 09:13:43.834693 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod55bb4fe2_94b0_4fdf_abb4_e39470fe2c30.slice/crio-56bc6b8828f391ce20802c03bbe556da08198fd0ad97378afbf331e3a28f1ad4 WatchSource:0}: Error finding container 56bc6b8828f391ce20802c03bbe556da08198fd0ad97378afbf331e3a28f1ad4: Status 404 returned error can't find the container with id 56bc6b8828f391ce20802c03bbe556da08198fd0ad97378afbf331e3a28f1ad4 Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.855715 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.855879 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwn8d\" (UniqueName: \"kubernetes.io/projected/37f3773a-2367-411b-b8e4-c4ae96970de3-kube-api-access-cwn8d\") pod \"control-plane-machine-set-operator-78cbb6b69f-j5djq\" (UID: \"37f3773a-2367-411b-b8e4-c4ae96970de3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5djq" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.855907 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d32fb60f-2d76-4e02-9839-535e1543eff7-config\") pod \"kube-apiserver-operator-766d6c64bb-s2hhg\" (UID: \"d32fb60f-2d76-4e02-9839-535e1543eff7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s2hhg" Mar 17 09:13:43 crc kubenswrapper[4813]: E0317 09:13:43.855952 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:44.355919561 +0000 UTC m=+246.456723060 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856007 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-oauth-serving-cert\") pod \"console-f9d7485db-2w9vb\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856066 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/ed67dcc9-2a14-4e28-86df-da77a8d20012-node-bootstrap-token\") pod \"machine-config-server-8hxbx\" (UID: \"ed67dcc9-2a14-4e28-86df-da77a8d20012\") " pod="openshift-machine-config-operator/machine-config-server-8hxbx" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856111 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7f82q\" (UniqueName: \"kubernetes.io/projected/0021cb4b-86a2-46a5-83f6-109fee3e517b-kube-api-access-7f82q\") pod \"olm-operator-6b444d44fb-tjvgc\" (UID: \"0021cb4b-86a2-46a5-83f6-109fee3e517b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tjvgc" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856140 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9902dff6-4484-44ba-af31-877293debb91-tmpfs\") pod \"packageserver-d55dfcdfc-slqwx\" (UID: \"9902dff6-4484-44ba-af31-877293debb91\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856163 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g44ww\" (UniqueName: \"kubernetes.io/projected/ed67dcc9-2a14-4e28-86df-da77a8d20012-kube-api-access-g44ww\") pod \"machine-config-server-8hxbx\" (UID: \"ed67dcc9-2a14-4e28-86df-da77a8d20012\") " pod="openshift-machine-config-operator/machine-config-server-8hxbx" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856188 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3992349-bc61-405e-be75-609cce2fad10-config-volume\") pod \"collect-profiles-29562300-vntp7\" (UID: \"b3992349-bc61-405e-be75-609cce2fad10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856213 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wslp6\" (UniqueName: \"kubernetes.io/projected/7954ca65-b7a4-47de-bc7c-2f57f0fee19a-kube-api-access-wslp6\") pod \"migrator-59844c95c7-tbl5t\" (UID: \"7954ca65-b7a4-47de-bc7c-2f57f0fee19a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tbl5t" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856243 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/782b4936-2caa-44da-8716-c57924be6df3-ca-trust-extracted\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856266 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/364cd6ac-033f-455e-a95a-10d50f055fb2-etcd-client\") pod \"etcd-operator-b45778765-jzbwp\" (UID: \"364cd6ac-033f-455e-a95a-10d50f055fb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856321 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/173f4a11-926b-4cc0-a70c-c369ac37bac2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6lsz5\" (UID: \"173f4a11-926b-4cc0-a70c-c369ac37bac2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6lsz5" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856352 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cnpcn\" (UniqueName: \"kubernetes.io/projected/845ee13b-7ed9-4910-958b-5269812237f0-kube-api-access-cnpcn\") pod \"cluster-samples-operator-665b6dd947-dpghh\" (UID: \"845ee13b-7ed9-4910-958b-5269812237f0\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dpghh" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856384 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d89e7afc-8f04-4996-b087-306543ef5acb-config\") pod \"service-ca-operator-777779d784-ghwhk\" (UID: \"d89e7afc-8f04-4996-b087-306543ef5acb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ghwhk" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856418 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/68399978-6f43-4458-a1e1-58d5984de8f5-stats-auth\") pod \"router-default-5444994796-m97th\" (UID: \"68399978-6f43-4458-a1e1-58d5984de8f5\") " pod="openshift-ingress/router-default-5444994796-m97th" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856451 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d32fb60f-2d76-4e02-9839-535e1543eff7-config\") pod \"kube-apiserver-operator-766d6c64bb-s2hhg\" (UID: \"d32fb60f-2d76-4e02-9839-535e1543eff7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s2hhg" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856452 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68399978-6f43-4458-a1e1-58d5984de8f5-metrics-certs\") pod \"router-default-5444994796-m97th\" (UID: \"68399978-6f43-4458-a1e1-58d5984de8f5\") " pod="openshift-ingress/router-default-5444994796-m97th" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856498 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72fa1c68-872b-45a3-98b7-b0b1ebc08ad7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wcd9z\" (UID: \"72fa1c68-872b-45a3-98b7-b0b1ebc08ad7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wcd9z" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856521 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/364cd6ac-033f-455e-a95a-10d50f055fb2-etcd-service-ca\") pod \"etcd-operator-b45778765-jzbwp\" (UID: \"364cd6ac-033f-455e-a95a-10d50f055fb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856537 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/364cd6ac-033f-455e-a95a-10d50f055fb2-config\") pod \"etcd-operator-b45778765-jzbwp\" (UID: \"364cd6ac-033f-455e-a95a-10d50f055fb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856555 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qr87f\" (UniqueName: \"kubernetes.io/projected/d89e7afc-8f04-4996-b087-306543ef5acb-kube-api-access-qr87f\") pod \"service-ca-operator-777779d784-ghwhk\" (UID: \"d89e7afc-8f04-4996-b087-306543ef5acb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ghwhk" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856575 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2bd4438d-8a23-4e8b-aad7-f5775f1d88a8-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wwbwr\" (UID: \"2bd4438d-8a23-4e8b-aad7-f5775f1d88a8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwbwr" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856617 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8745x\" (UniqueName: \"kubernetes.io/projected/93cb7b2d-1c40-40be-82d7-1858272444ec-kube-api-access-8745x\") pod \"cluster-image-registry-operator-dc59b4c8b-njr98\" (UID: \"93cb7b2d-1c40-40be-82d7-1858272444ec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-njr98" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856643 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcghd\" (UniqueName: \"kubernetes.io/projected/8626af43-6b5e-49e1-b19f-3986cdecdd2d-kube-api-access-fcghd\") pod \"auto-csr-approver-29562312-hjpcp\" (UID: \"8626af43-6b5e-49e1-b19f-3986cdecdd2d\") " pod="openshift-infra/auto-csr-approver-29562312-hjpcp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856665 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/cfccf8af-20ed-4d82-96d7-b4b3c00cd24b-socket-dir\") pod \"csi-hostpathplugin-gczkp\" (UID: \"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b\") " pod="hostpath-provisioner/csi-hostpathplugin-gczkp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856684 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b828a0e2-3def-4bb5-8f54-2792376423cd-serving-cert\") pod \"openshift-config-operator-7777fb866f-4rqzj\" (UID: \"b828a0e2-3def-4bb5-8f54-2792376423cd\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4rqzj" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856701 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6564\" (UniqueName: \"kubernetes.io/projected/da36625d-0c68-48b8-9608-c7dedd304064-kube-api-access-h6564\") pod \"dns-default-cjn8k\" (UID: \"da36625d-0c68-48b8-9608-c7dedd304064\") " pod="openshift-dns/dns-default-cjn8k" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856716 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0021cb4b-86a2-46a5-83f6-109fee3e517b-srv-cert\") pod \"olm-operator-6b444d44fb-tjvgc\" (UID: \"0021cb4b-86a2-46a5-83f6-109fee3e517b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tjvgc" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856733 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kpqgd\" (UniqueName: \"kubernetes.io/projected/a6f67978-6c47-4927-ae94-1fef4eee7b18-kube-api-access-kpqgd\") pod \"ingress-canary-rw942\" (UID: \"a6f67978-6c47-4927-ae94-1fef4eee7b18\") " pod="openshift-ingress-canary/ingress-canary-rw942" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856748 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/58e4f2e1-b751-41e1-a56e-b62fd82b7caf-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-7tdh4\" (UID: \"58e4f2e1-b751-41e1-a56e-b62fd82b7caf\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7tdh4" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856781 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2657\" (UniqueName: \"kubernetes.io/projected/550c18c5-f4e0-4da1-9692-c6cacfa882af-kube-api-access-z2657\") pod \"authentication-operator-69f744f599-tzt6n\" (UID: \"550c18c5-f4e0-4da1-9692-c6cacfa882af\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tzt6n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856805 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/94f282fa-7b22-41e3-9911-073b05d41200-metrics-tls\") pod \"dns-operator-744455d44c-qwlbl\" (UID: \"94f282fa-7b22-41e3-9911-073b05d41200\") " pod="openshift-dns-operator/dns-operator-744455d44c-qwlbl" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856826 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9902dff6-4484-44ba-af31-877293debb91-webhook-cert\") pod \"packageserver-d55dfcdfc-slqwx\" (UID: \"9902dff6-4484-44ba-af31-877293debb91\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856850 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72fa1c68-872b-45a3-98b7-b0b1ebc08ad7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wcd9z\" (UID: \"72fa1c68-872b-45a3-98b7-b0b1ebc08ad7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wcd9z" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856874 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/58e4f2e1-b751-41e1-a56e-b62fd82b7caf-proxy-tls\") pod \"machine-config-controller-84d6567774-7tdh4\" (UID: \"58e4f2e1-b751-41e1-a56e-b62fd82b7caf\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7tdh4" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856895 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vg88g\" (UniqueName: \"kubernetes.io/projected/f760de2f-292d-48fe-ac0c-a97aa37f556a-kube-api-access-vg88g\") pod \"catalog-operator-68c6474976-8qg9n\" (UID: \"f760de2f-292d-48fe-ac0c-a97aa37f556a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8qg9n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856922 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/782b4936-2caa-44da-8716-c57924be6df3-bound-sa-token\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856947 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74f92156-f803-433a-99a4-981202076e81-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5wl9g\" (UID: \"74f92156-f803-433a-99a4-981202076e81\") " pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856967 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0021cb4b-86a2-46a5-83f6-109fee3e517b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-tjvgc\" (UID: \"0021cb4b-86a2-46a5-83f6-109fee3e517b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tjvgc" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.856998 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/93cb7b2d-1c40-40be-82d7-1858272444ec-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-njr98\" (UID: \"93cb7b2d-1c40-40be-82d7-1858272444ec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-njr98" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857017 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/550c18c5-f4e0-4da1-9692-c6cacfa882af-serving-cert\") pod \"authentication-operator-69f744f599-tzt6n\" (UID: \"550c18c5-f4e0-4da1-9692-c6cacfa882af\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tzt6n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857053 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/782b4936-2caa-44da-8716-c57924be6df3-trusted-ca\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857076 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zgwpz\" (UniqueName: \"kubernetes.io/projected/000e63dc-8228-486c-a6ee-12843810b5af-kube-api-access-zgwpz\") pod \"service-ca-9c57cc56f-wttgv\" (UID: \"000e63dc-8228-486c-a6ee-12843810b5af\") " pod="openshift-service-ca/service-ca-9c57cc56f-wttgv" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857092 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f7c4987b-f8f6-4d39-b7f1-6f80c1384293-metrics-tls\") pod \"ingress-operator-5b745b69d9-s6zwp\" (UID: \"f7c4987b-f8f6-4d39-b7f1-6f80c1384293\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s6zwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857110 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/845ee13b-7ed9-4910-958b-5269812237f0-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-dpghh\" (UID: \"845ee13b-7ed9-4910-958b-5269812237f0\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dpghh" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857125 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/364cd6ac-033f-455e-a95a-10d50f055fb2-serving-cert\") pod \"etcd-operator-b45778765-jzbwp\" (UID: \"364cd6ac-033f-455e-a95a-10d50f055fb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857143 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/cfccf8af-20ed-4d82-96d7-b4b3c00cd24b-plugins-dir\") pod \"csi-hostpathplugin-gczkp\" (UID: \"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b\") " pod="hostpath-provisioner/csi-hostpathplugin-gczkp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857158 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03c2477b-c56a-4354-bf4f-170b83f9240f-config\") pod \"kube-controller-manager-operator-78b949d7b-2bpfb\" (UID: \"03c2477b-c56a-4354-bf4f-170b83f9240f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bpfb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857187 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857204 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f7603094-bb1d-4d77-b463-2270003b2805-console-oauth-config\") pod \"console-f9d7485db-2w9vb\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857221 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8hk6\" (UniqueName: \"kubernetes.io/projected/74f92156-f803-433a-99a4-981202076e81-kube-api-access-q8hk6\") pod \"marketplace-operator-79b997595-5wl9g\" (UID: \"74f92156-f803-433a-99a4-981202076e81\") " pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857247 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f2q6\" (UniqueName: \"kubernetes.io/projected/94f282fa-7b22-41e3-9911-073b05d41200-kube-api-access-7f2q6\") pod \"dns-operator-744455d44c-qwlbl\" (UID: \"94f282fa-7b22-41e3-9911-073b05d41200\") " pod="openshift-dns-operator/dns-operator-744455d44c-qwlbl" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857271 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/782b4936-2caa-44da-8716-c57924be6df3-installation-pull-secrets\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857300 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2bd4438d-8a23-4e8b-aad7-f5775f1d88a8-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wwbwr\" (UID: \"2bd4438d-8a23-4e8b-aad7-f5775f1d88a8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwbwr" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857317 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nk8pf\" (UniqueName: \"kubernetes.io/projected/1507ff6e-1511-4c90-998e-1fb84da97e5b-kube-api-access-nk8pf\") pod \"machine-config-operator-74547568cd-vm474\" (UID: \"1507ff6e-1511-4c90-998e-1fb84da97e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vm474" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857335 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/782b4936-2caa-44da-8716-c57924be6df3-registry-certificates\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857353 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3abdbcf2-c3d0-4049-b08b-45fa02672dd4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-7w9l5\" (UID: \"3abdbcf2-c3d0-4049-b08b-45fa02672dd4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7w9l5" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857369 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/93cb7b2d-1c40-40be-82d7-1858272444ec-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-njr98\" (UID: \"93cb7b2d-1c40-40be-82d7-1858272444ec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-njr98" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857386 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f760de2f-292d-48fe-ac0c-a97aa37f556a-srv-cert\") pod \"catalog-operator-68c6474976-8qg9n\" (UID: \"f760de2f-292d-48fe-ac0c-a97aa37f556a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8qg9n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857413 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/68399978-6f43-4458-a1e1-58d5984de8f5-service-ca-bundle\") pod \"router-default-5444994796-m97th\" (UID: \"68399978-6f43-4458-a1e1-58d5984de8f5\") " pod="openshift-ingress/router-default-5444994796-m97th" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857428 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f7c4987b-f8f6-4d39-b7f1-6f80c1384293-bound-sa-token\") pod \"ingress-operator-5b745b69d9-s6zwp\" (UID: \"f7c4987b-f8f6-4d39-b7f1-6f80c1384293\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s6zwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857445 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d89e7afc-8f04-4996-b087-306543ef5acb-serving-cert\") pod \"service-ca-operator-777779d784-ghwhk\" (UID: \"d89e7afc-8f04-4996-b087-306543ef5acb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ghwhk" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857471 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdnwc\" (UniqueName: \"kubernetes.io/projected/782b4936-2caa-44da-8716-c57924be6df3-kube-api-access-kdnwc\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857486 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/74f92156-f803-433a-99a4-981202076e81-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5wl9g\" (UID: \"74f92156-f803-433a-99a4-981202076e81\") " pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857510 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f7603094-bb1d-4d77-b463-2270003b2805-console-serving-cert\") pod \"console-f9d7485db-2w9vb\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857526 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlzt2\" (UniqueName: \"kubernetes.io/projected/58e4f2e1-b751-41e1-a56e-b62fd82b7caf-kube-api-access-tlzt2\") pod \"machine-config-controller-84d6567774-7tdh4\" (UID: \"58e4f2e1-b751-41e1-a56e-b62fd82b7caf\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7tdh4" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857556 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24qg4\" (UniqueName: \"kubernetes.io/projected/3aa0fc32-ed11-482b-9a6d-fcd59544bf8d-kube-api-access-24qg4\") pod \"downloads-7954f5f757-xb4k9\" (UID: \"3aa0fc32-ed11-482b-9a6d-fcd59544bf8d\") " pod="openshift-console/downloads-7954f5f757-xb4k9" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857573 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwnf4\" (UniqueName: \"kubernetes.io/projected/9902dff6-4484-44ba-af31-877293debb91-kube-api-access-qwnf4\") pod \"packageserver-d55dfcdfc-slqwx\" (UID: \"9902dff6-4484-44ba-af31-877293debb91\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857617 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a6f67978-6c47-4927-ae94-1fef4eee7b18-cert\") pod \"ingress-canary-rw942\" (UID: \"a6f67978-6c47-4927-ae94-1fef4eee7b18\") " pod="openshift-ingress-canary/ingress-canary-rw942" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857639 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4lhgk\" (UniqueName: \"kubernetes.io/projected/68399978-6f43-4458-a1e1-58d5984de8f5-kube-api-access-4lhgk\") pod \"router-default-5444994796-m97th\" (UID: \"68399978-6f43-4458-a1e1-58d5984de8f5\") " pod="openshift-ingress/router-default-5444994796-m97th" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857663 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/cfccf8af-20ed-4d82-96d7-b4b3c00cd24b-mountpoint-dir\") pod \"csi-hostpathplugin-gczkp\" (UID: \"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b\") " pod="hostpath-provisioner/csi-hostpathplugin-gczkp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857688 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12efeac2-9fc3-425a-8874-3c2d957e04ca-config\") pod \"machine-approver-56656f9798-kpxlg\" (UID: \"12efeac2-9fc3-425a-8874-3c2d957e04ca\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kpxlg" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857701 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/cfccf8af-20ed-4d82-96d7-b4b3c00cd24b-registration-dir\") pod \"csi-hostpathplugin-gczkp\" (UID: \"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b\") " pod="hostpath-provisioner/csi-hostpathplugin-gczkp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857717 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/cfccf8af-20ed-4d82-96d7-b4b3c00cd24b-csi-data-dir\") pod \"csi-hostpathplugin-gczkp\" (UID: \"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b\") " pod="hostpath-provisioner/csi-hostpathplugin-gczkp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857734 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n26g7\" (UniqueName: \"kubernetes.io/projected/b3992349-bc61-405e-be75-609cce2fad10-kube-api-access-n26g7\") pod \"collect-profiles-29562300-vntp7\" (UID: \"b3992349-bc61-405e-be75-609cce2fad10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857751 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1507ff6e-1511-4c90-998e-1fb84da97e5b-proxy-tls\") pod \"machine-config-operator-74547568cd-vm474\" (UID: \"1507ff6e-1511-4c90-998e-1fb84da97e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vm474" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857795 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03c2477b-c56a-4354-bf4f-170b83f9240f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-2bpfb\" (UID: \"03c2477b-c56a-4354-bf4f-170b83f9240f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bpfb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857813 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dx7gs\" (UniqueName: \"kubernetes.io/projected/364cd6ac-033f-455e-a95a-10d50f055fb2-kube-api-access-dx7gs\") pod \"etcd-operator-b45778765-jzbwp\" (UID: \"364cd6ac-033f-455e-a95a-10d50f055fb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857828 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9902dff6-4484-44ba-af31-877293debb91-apiservice-cert\") pod \"packageserver-d55dfcdfc-slqwx\" (UID: \"9902dff6-4484-44ba-af31-877293debb91\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857847 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f253587f-9f27-474d-8f46-21f1e6a79d97-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-lrhrs\" (UID: \"f253587f-9f27-474d-8f46-21f1e6a79d97\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lrhrs" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857878 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/550c18c5-f4e0-4da1-9692-c6cacfa882af-config\") pod \"authentication-operator-69f744f599-tzt6n\" (UID: \"550c18c5-f4e0-4da1-9692-c6cacfa882af\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tzt6n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857905 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-oauth-serving-cert\") pod \"console-f9d7485db-2w9vb\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857935 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bd4438d-8a23-4e8b-aad7-f5775f1d88a8-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wwbwr\" (UID: \"2bd4438d-8a23-4e8b-aad7-f5775f1d88a8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwbwr" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.859230 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5x5b4\" (UniqueName: \"kubernetes.io/projected/f7c4987b-f8f6-4d39-b7f1-6f80c1384293-kube-api-access-5x5b4\") pod \"ingress-operator-5b745b69d9-s6zwp\" (UID: \"f7c4987b-f8f6-4d39-b7f1-6f80c1384293\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s6zwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.859269 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z872t\" (UniqueName: \"kubernetes.io/projected/cfccf8af-20ed-4d82-96d7-b4b3c00cd24b-kube-api-access-z872t\") pod \"csi-hostpathplugin-gczkp\" (UID: \"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b\") " pod="hostpath-provisioner/csi-hostpathplugin-gczkp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.859286 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/000e63dc-8228-486c-a6ee-12843810b5af-signing-cabundle\") pod \"service-ca-9c57cc56f-wttgv\" (UID: \"000e63dc-8228-486c-a6ee-12843810b5af\") " pod="openshift-service-ca/service-ca-9c57cc56f-wttgv" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.859303 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/da36625d-0c68-48b8-9608-c7dedd304064-metrics-tls\") pod \"dns-default-cjn8k\" (UID: \"da36625d-0c68-48b8-9608-c7dedd304064\") " pod="openshift-dns/dns-default-cjn8k" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.859320 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1507ff6e-1511-4c90-998e-1fb84da97e5b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vm474\" (UID: \"1507ff6e-1511-4c90-998e-1fb84da97e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vm474" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.859338 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/37f3773a-2367-411b-b8e4-c4ae96970de3-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-j5djq\" (UID: \"37f3773a-2367-411b-b8e4-c4ae96970de3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5djq" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.859372 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/782b4936-2caa-44da-8716-c57924be6df3-registry-tls\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.859390 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/12efeac2-9fc3-425a-8874-3c2d957e04ca-auth-proxy-config\") pod \"machine-approver-56656f9798-kpxlg\" (UID: \"12efeac2-9fc3-425a-8874-3c2d957e04ca\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kpxlg" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.859933 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3992349-bc61-405e-be75-609cce2fad10-secret-volume\") pod \"collect-profiles-29562300-vntp7\" (UID: \"b3992349-bc61-405e-be75-609cce2fad10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.859965 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq869\" (UniqueName: \"kubernetes.io/projected/173f4a11-926b-4cc0-a70c-c369ac37bac2-kube-api-access-pq869\") pod \"multus-admission-controller-857f4d67dd-6lsz5\" (UID: \"173f4a11-926b-4cc0-a70c-c369ac37bac2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6lsz5" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860056 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/03c2477b-c56a-4354-bf4f-170b83f9240f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-2bpfb\" (UID: \"03c2477b-c56a-4354-bf4f-170b83f9240f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bpfb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860074 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/000e63dc-8228-486c-a6ee-12843810b5af-signing-key\") pod \"service-ca-9c57cc56f-wttgv\" (UID: \"000e63dc-8228-486c-a6ee-12843810b5af\") " pod="openshift-service-ca/service-ca-9c57cc56f-wttgv" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860097 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjfgf\" (UniqueName: \"kubernetes.io/projected/b828a0e2-3def-4bb5-8f54-2792376423cd-kube-api-access-xjfgf\") pod \"openshift-config-operator-7777fb866f-4rqzj\" (UID: \"b828a0e2-3def-4bb5-8f54-2792376423cd\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4rqzj" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860117 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-trusted-ca-bundle\") pod \"console-f9d7485db-2w9vb\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860138 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f7c4987b-f8f6-4d39-b7f1-6f80c1384293-trusted-ca\") pod \"ingress-operator-5b745b69d9-s6zwp\" (UID: \"f7c4987b-f8f6-4d39-b7f1-6f80c1384293\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s6zwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860161 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/550c18c5-f4e0-4da1-9692-c6cacfa882af-service-ca-bundle\") pod \"authentication-operator-69f744f599-tzt6n\" (UID: \"550c18c5-f4e0-4da1-9692-c6cacfa882af\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tzt6n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860180 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d32fb60f-2d76-4e02-9839-535e1543eff7-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-s2hhg\" (UID: \"d32fb60f-2d76-4e02-9839-535e1543eff7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s2hhg" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860202 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b828a0e2-3def-4bb5-8f54-2792376423cd-available-featuregates\") pod \"openshift-config-operator-7777fb866f-4rqzj\" (UID: \"b828a0e2-3def-4bb5-8f54-2792376423cd\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4rqzj" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860220 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d32fb60f-2d76-4e02-9839-535e1543eff7-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-s2hhg\" (UID: \"d32fb60f-2d76-4e02-9839-535e1543eff7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s2hhg" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860236 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f760de2f-292d-48fe-ac0c-a97aa37f556a-profile-collector-cert\") pod \"catalog-operator-68c6474976-8qg9n\" (UID: \"f760de2f-292d-48fe-ac0c-a97aa37f556a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8qg9n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860275 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/ed67dcc9-2a14-4e28-86df-da77a8d20012-certs\") pod \"machine-config-server-8hxbx\" (UID: \"ed67dcc9-2a14-4e28-86df-da77a8d20012\") " pod="openshift-machine-config-operator/machine-config-server-8hxbx" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860293 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ds7cb\" (UniqueName: \"kubernetes.io/projected/3abdbcf2-c3d0-4049-b08b-45fa02672dd4-kube-api-access-ds7cb\") pod \"package-server-manager-789f6589d5-7w9l5\" (UID: \"3abdbcf2-c3d0-4049-b08b-45fa02672dd4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7w9l5" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860315 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s29cv\" (UniqueName: \"kubernetes.io/projected/12efeac2-9fc3-425a-8874-3c2d957e04ca-kube-api-access-s29cv\") pod \"machine-approver-56656f9798-kpxlg\" (UID: \"12efeac2-9fc3-425a-8874-3c2d957e04ca\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kpxlg" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860332 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/93cb7b2d-1c40-40be-82d7-1858272444ec-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-njr98\" (UID: \"93cb7b2d-1c40-40be-82d7-1858272444ec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-njr98" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860348 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/364cd6ac-033f-455e-a95a-10d50f055fb2-etcd-ca\") pod \"etcd-operator-b45778765-jzbwp\" (UID: \"364cd6ac-033f-455e-a95a-10d50f055fb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860376 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-console-config\") pod \"console-f9d7485db-2w9vb\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860392 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-service-ca\") pod \"console-f9d7485db-2w9vb\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860410 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1507ff6e-1511-4c90-998e-1fb84da97e5b-images\") pod \"machine-config-operator-74547568cd-vm474\" (UID: \"1507ff6e-1511-4c90-998e-1fb84da97e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vm474" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860438 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/550c18c5-f4e0-4da1-9692-c6cacfa882af-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-tzt6n\" (UID: \"550c18c5-f4e0-4da1-9692-c6cacfa882af\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tzt6n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860457 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/12efeac2-9fc3-425a-8874-3c2d957e04ca-machine-approver-tls\") pod \"machine-approver-56656f9798-kpxlg\" (UID: \"12efeac2-9fc3-425a-8874-3c2d957e04ca\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kpxlg" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860477 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/68399978-6f43-4458-a1e1-58d5984de8f5-default-certificate\") pod \"router-default-5444994796-m97th\" (UID: \"68399978-6f43-4458-a1e1-58d5984de8f5\") " pod="openshift-ingress/router-default-5444994796-m97th" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860499 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzrsb\" (UniqueName: \"kubernetes.io/projected/72fa1c68-872b-45a3-98b7-b0b1ebc08ad7-kube-api-access-tzrsb\") pod \"openshift-controller-manager-operator-756b6f6bc6-wcd9z\" (UID: \"72fa1c68-872b-45a3-98b7-b0b1ebc08ad7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wcd9z" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860515 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f253587f-9f27-474d-8f46-21f1e6a79d97-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-lrhrs\" (UID: \"f253587f-9f27-474d-8f46-21f1e6a79d97\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lrhrs" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860538 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngkkm\" (UniqueName: \"kubernetes.io/projected/f7603094-bb1d-4d77-b463-2270003b2805-kube-api-access-ngkkm\") pod \"console-f9d7485db-2w9vb\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860544 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/364cd6ac-033f-455e-a95a-10d50f055fb2-etcd-client\") pod \"etcd-operator-b45778765-jzbwp\" (UID: \"364cd6ac-033f-455e-a95a-10d50f055fb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860573 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da36625d-0c68-48b8-9608-c7dedd304064-config-volume\") pod \"dns-default-cjn8k\" (UID: \"da36625d-0c68-48b8-9608-c7dedd304064\") " pod="openshift-dns/dns-default-cjn8k" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.860610 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gssm9\" (UniqueName: \"kubernetes.io/projected/f253587f-9f27-474d-8f46-21f1e6a79d97-kube-api-access-gssm9\") pod \"kube-storage-version-migrator-operator-b67b599dd-lrhrs\" (UID: \"f253587f-9f27-474d-8f46-21f1e6a79d97\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lrhrs" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.857808 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/782b4936-2caa-44da-8716-c57924be6df3-ca-trust-extracted\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.861663 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/68399978-6f43-4458-a1e1-58d5984de8f5-metrics-certs\") pod \"router-default-5444994796-m97th\" (UID: \"68399978-6f43-4458-a1e1-58d5984de8f5\") " pod="openshift-ingress/router-default-5444994796-m97th" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.862179 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f7603094-bb1d-4d77-b463-2270003b2805-console-serving-cert\") pod \"console-f9d7485db-2w9vb\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.862862 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2bd4438d-8a23-4e8b-aad7-f5775f1d88a8-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wwbwr\" (UID: \"2bd4438d-8a23-4e8b-aad7-f5775f1d88a8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwbwr" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.863280 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/72fa1c68-872b-45a3-98b7-b0b1ebc08ad7-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-wcd9z\" (UID: \"72fa1c68-872b-45a3-98b7-b0b1ebc08ad7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wcd9z" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.863464 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/550c18c5-f4e0-4da1-9692-c6cacfa882af-service-ca-bundle\") pod \"authentication-operator-69f744f599-tzt6n\" (UID: \"550c18c5-f4e0-4da1-9692-c6cacfa882af\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tzt6n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.863774 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/b828a0e2-3def-4bb5-8f54-2792376423cd-available-featuregates\") pod \"openshift-config-operator-7777fb866f-4rqzj\" (UID: \"b828a0e2-3def-4bb5-8f54-2792376423cd\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4rqzj" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.864775 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/12efeac2-9fc3-425a-8874-3c2d957e04ca-auth-proxy-config\") pod \"machine-approver-56656f9798-kpxlg\" (UID: \"12efeac2-9fc3-425a-8874-3c2d957e04ca\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kpxlg" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.864883 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12efeac2-9fc3-425a-8874-3c2d957e04ca-config\") pod \"machine-approver-56656f9798-kpxlg\" (UID: \"12efeac2-9fc3-425a-8874-3c2d957e04ca\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kpxlg" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.865777 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/550c18c5-f4e0-4da1-9692-c6cacfa882af-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-tzt6n\" (UID: \"550c18c5-f4e0-4da1-9692-c6cacfa882af\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tzt6n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.866811 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-service-ca\") pod \"console-f9d7485db-2w9vb\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.866821 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/93cb7b2d-1c40-40be-82d7-1858272444ec-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-njr98\" (UID: \"93cb7b2d-1c40-40be-82d7-1858272444ec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-njr98" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.866933 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-trusted-ca-bundle\") pod \"console-f9d7485db-2w9vb\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.867144 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-console-config\") pod \"console-f9d7485db-2w9vb\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.867447 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/364cd6ac-033f-455e-a95a-10d50f055fb2-etcd-ca\") pod \"etcd-operator-b45778765-jzbwp\" (UID: \"364cd6ac-033f-455e-a95a-10d50f055fb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" Mar 17 09:13:43 crc kubenswrapper[4813]: E0317 09:13:43.867843 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:44.367821337 +0000 UTC m=+246.468624836 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.868009 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/68399978-6f43-4458-a1e1-58d5984de8f5-service-ca-bundle\") pod \"router-default-5444994796-m97th\" (UID: \"68399978-6f43-4458-a1e1-58d5984de8f5\") " pod="openshift-ingress/router-default-5444994796-m97th" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.868776 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/550c18c5-f4e0-4da1-9692-c6cacfa882af-config\") pod \"authentication-operator-69f744f599-tzt6n\" (UID: \"550c18c5-f4e0-4da1-9692-c6cacfa882af\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tzt6n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.869450 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/72fa1c68-872b-45a3-98b7-b0b1ebc08ad7-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-wcd9z\" (UID: \"72fa1c68-872b-45a3-98b7-b0b1ebc08ad7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wcd9z" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.869586 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/782b4936-2caa-44da-8716-c57924be6df3-registry-certificates\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.869764 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/364cd6ac-033f-455e-a95a-10d50f055fb2-config\") pod \"etcd-operator-b45778765-jzbwp\" (UID: \"364cd6ac-033f-455e-a95a-10d50f055fb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.870135 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f7603094-bb1d-4d77-b463-2270003b2805-console-oauth-config\") pod \"console-f9d7485db-2w9vb\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.870713 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/550c18c5-f4e0-4da1-9692-c6cacfa882af-serving-cert\") pod \"authentication-operator-69f744f599-tzt6n\" (UID: \"550c18c5-f4e0-4da1-9692-c6cacfa882af\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tzt6n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.870744 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/364cd6ac-033f-455e-a95a-10d50f055fb2-etcd-service-ca\") pod \"etcd-operator-b45778765-jzbwp\" (UID: \"364cd6ac-033f-455e-a95a-10d50f055fb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.871061 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d32fb60f-2d76-4e02-9839-535e1543eff7-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-s2hhg\" (UID: \"d32fb60f-2d76-4e02-9839-535e1543eff7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s2hhg" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.871146 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/94f282fa-7b22-41e3-9911-073b05d41200-metrics-tls\") pod \"dns-operator-744455d44c-qwlbl\" (UID: \"94f282fa-7b22-41e3-9911-073b05d41200\") " pod="openshift-dns-operator/dns-operator-744455d44c-qwlbl" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.871236 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/68399978-6f43-4458-a1e1-58d5984de8f5-default-certificate\") pod \"router-default-5444994796-m97th\" (UID: \"68399978-6f43-4458-a1e1-58d5984de8f5\") " pod="openshift-ingress/router-default-5444994796-m97th" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.871522 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/782b4936-2caa-44da-8716-c57924be6df3-installation-pull-secrets\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.871873 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/845ee13b-7ed9-4910-958b-5269812237f0-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-dpghh\" (UID: \"845ee13b-7ed9-4910-958b-5269812237f0\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dpghh" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.871987 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b828a0e2-3def-4bb5-8f54-2792376423cd-serving-cert\") pod \"openshift-config-operator-7777fb866f-4rqzj\" (UID: \"b828a0e2-3def-4bb5-8f54-2792376423cd\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4rqzj" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.873282 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/782b4936-2caa-44da-8716-c57924be6df3-trusted-ca\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.873854 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/93cb7b2d-1c40-40be-82d7-1858272444ec-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-njr98\" (UID: \"93cb7b2d-1c40-40be-82d7-1858272444ec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-njr98" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.874029 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/782b4936-2caa-44da-8716-c57924be6df3-registry-tls\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.874156 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/12efeac2-9fc3-425a-8874-3c2d957e04ca-machine-approver-tls\") pod \"machine-approver-56656f9798-kpxlg\" (UID: \"12efeac2-9fc3-425a-8874-3c2d957e04ca\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kpxlg" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.875248 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2bd4438d-8a23-4e8b-aad7-f5775f1d88a8-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wwbwr\" (UID: \"2bd4438d-8a23-4e8b-aad7-f5775f1d88a8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwbwr" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.875454 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/68399978-6f43-4458-a1e1-58d5984de8f5-stats-auth\") pod \"router-default-5444994796-m97th\" (UID: \"68399978-6f43-4458-a1e1-58d5984de8f5\") " pod="openshift-ingress/router-default-5444994796-m97th" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.876258 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/364cd6ac-033f-455e-a95a-10d50f055fb2-serving-cert\") pod \"etcd-operator-b45778765-jzbwp\" (UID: \"364cd6ac-033f-455e-a95a-10d50f055fb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.886481 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdnwc\" (UniqueName: \"kubernetes.io/projected/782b4936-2caa-44da-8716-c57924be6df3-kube-api-access-kdnwc\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.898630 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm"] Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.909292 4813 generic.go:334] "Generic (PLEG): container finished" podID="26a86c9b-1645-49c2-a2fb-78b210984363" containerID="9f98ce818acce7cbad788134699f1b5fe1b0be4690685e0bb1a272109e717050" exitCode=0 Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.909367 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kns96" event={"ID":"26a86c9b-1645-49c2-a2fb-78b210984363","Type":"ContainerDied","Data":"9f98ce818acce7cbad788134699f1b5fe1b0be4690685e0bb1a272109e717050"} Mar 17 09:13:43 crc kubenswrapper[4813]: W0317 09:13:43.910561 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podeb147d37_445a_48f2_a2e6_fb16f3575529.slice/crio-95e77e3f1afe36af208b6ee8400904d1eb126d3f625a2467bd94cc39335dc94a WatchSource:0}: Error finding container 95e77e3f1afe36af208b6ee8400904d1eb126d3f625a2467bd94cc39335dc94a: Status 404 returned error can't find the container with id 95e77e3f1afe36af208b6ee8400904d1eb126d3f625a2467bd94cc39335dc94a Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.916822 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5kp7w"] Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.918726 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-sk47b" event={"ID":"feae3098-e93c-4dff-9088-a38db6a5d929","Type":"ContainerStarted","Data":"32b5327e411d4faa26a001a54967917b9f75b9e84aebb536dec47ed6ea33bf62"} Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.919046 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/782b4936-2caa-44da-8716-c57924be6df3-bound-sa-token\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.921929 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" event={"ID":"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30","Type":"ContainerStarted","Data":"56bc6b8828f391ce20802c03bbe556da08198fd0ad97378afbf331e3a28f1ad4"} Mar 17 09:13:43 crc kubenswrapper[4813]: W0317 09:13:43.926984 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e837d92_c48e_4b0f_a55a_40bdaea5d9f0.slice/crio-0a9009f31556196a712dbea20111012eb1c927a32e77a31daaae2fc5e415da10 WatchSource:0}: Error finding container 0a9009f31556196a712dbea20111012eb1c927a32e77a31daaae2fc5e415da10: Status 404 returned error can't find the container with id 0a9009f31556196a712dbea20111012eb1c927a32e77a31daaae2fc5e415da10 Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.934524 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k58b"] Mar 17 09:13:43 crc kubenswrapper[4813]: W0317 09:13:43.940729 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod98211622_4794_4412_a9d0_64436ee1a3f4.slice/crio-505f761af0bd804e9abeacc49ac4f0c6abc151024c424e6073a665c371883f06 WatchSource:0}: Error finding container 505f761af0bd804e9abeacc49ac4f0c6abc151024c424e6073a665c371883f06: Status 404 returned error can't find the container with id 505f761af0bd804e9abeacc49ac4f0c6abc151024c424e6073a665c371883f06 Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.942262 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjfgf\" (UniqueName: \"kubernetes.io/projected/b828a0e2-3def-4bb5-8f54-2792376423cd-kube-api-access-xjfgf\") pod \"openshift-config-operator-7777fb866f-4rqzj\" (UID: \"b828a0e2-3def-4bb5-8f54-2792376423cd\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-4rqzj" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.949444 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-2fmh2"] Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.959646 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/93cb7b2d-1c40-40be-82d7-1858272444ec-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-njr98\" (UID: \"93cb7b2d-1c40-40be-82d7-1858272444ec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-njr98" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.962419 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:43 crc kubenswrapper[4813]: E0317 09:13:43.962644 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:44.462561463 +0000 UTC m=+246.563364972 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.962699 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zgwpz\" (UniqueName: \"kubernetes.io/projected/000e63dc-8228-486c-a6ee-12843810b5af-kube-api-access-zgwpz\") pod \"service-ca-9c57cc56f-wttgv\" (UID: \"000e63dc-8228-486c-a6ee-12843810b5af\") " pod="openshift-service-ca/service-ca-9c57cc56f-wttgv" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.962740 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f7c4987b-f8f6-4d39-b7f1-6f80c1384293-metrics-tls\") pod \"ingress-operator-5b745b69d9-s6zwp\" (UID: \"f7c4987b-f8f6-4d39-b7f1-6f80c1384293\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s6zwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.962767 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/cfccf8af-20ed-4d82-96d7-b4b3c00cd24b-plugins-dir\") pod \"csi-hostpathplugin-gczkp\" (UID: \"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b\") " pod="hostpath-provisioner/csi-hostpathplugin-gczkp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.962788 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03c2477b-c56a-4354-bf4f-170b83f9240f-config\") pod \"kube-controller-manager-operator-78b949d7b-2bpfb\" (UID: \"03c2477b-c56a-4354-bf4f-170b83f9240f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bpfb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.962818 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.962883 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8hk6\" (UniqueName: \"kubernetes.io/projected/74f92156-f803-433a-99a4-981202076e81-kube-api-access-q8hk6\") pod \"marketplace-operator-79b997595-5wl9g\" (UID: \"74f92156-f803-433a-99a4-981202076e81\") " pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.962952 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nk8pf\" (UniqueName: \"kubernetes.io/projected/1507ff6e-1511-4c90-998e-1fb84da97e5b-kube-api-access-nk8pf\") pod \"machine-config-operator-74547568cd-vm474\" (UID: \"1507ff6e-1511-4c90-998e-1fb84da97e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vm474" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.962978 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3abdbcf2-c3d0-4049-b08b-45fa02672dd4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-7w9l5\" (UID: \"3abdbcf2-c3d0-4049-b08b-45fa02672dd4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7w9l5" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.963002 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f760de2f-292d-48fe-ac0c-a97aa37f556a-srv-cert\") pod \"catalog-operator-68c6474976-8qg9n\" (UID: \"f760de2f-292d-48fe-ac0c-a97aa37f556a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8qg9n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.963029 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f7c4987b-f8f6-4d39-b7f1-6f80c1384293-bound-sa-token\") pod \"ingress-operator-5b745b69d9-s6zwp\" (UID: \"f7c4987b-f8f6-4d39-b7f1-6f80c1384293\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s6zwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.963051 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d89e7afc-8f04-4996-b087-306543ef5acb-serving-cert\") pod \"service-ca-operator-777779d784-ghwhk\" (UID: \"d89e7afc-8f04-4996-b087-306543ef5acb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ghwhk" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.963076 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/74f92156-f803-433a-99a4-981202076e81-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5wl9g\" (UID: \"74f92156-f803-433a-99a4-981202076e81\") " pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.963111 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlzt2\" (UniqueName: \"kubernetes.io/projected/58e4f2e1-b751-41e1-a56e-b62fd82b7caf-kube-api-access-tlzt2\") pod \"machine-config-controller-84d6567774-7tdh4\" (UID: \"58e4f2e1-b751-41e1-a56e-b62fd82b7caf\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7tdh4" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.963279 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/cfccf8af-20ed-4d82-96d7-b4b3c00cd24b-plugins-dir\") pod \"csi-hostpathplugin-gczkp\" (UID: \"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b\") " pod="hostpath-provisioner/csi-hostpathplugin-gczkp" Mar 17 09:13:43 crc kubenswrapper[4813]: E0317 09:13:43.963410 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:44.463399642 +0000 UTC m=+246.564203141 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.963861 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/03c2477b-c56a-4354-bf4f-170b83f9240f-config\") pod \"kube-controller-manager-operator-78b949d7b-2bpfb\" (UID: \"03c2477b-c56a-4354-bf4f-170b83f9240f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bpfb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965170 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwnf4\" (UniqueName: \"kubernetes.io/projected/9902dff6-4484-44ba-af31-877293debb91-kube-api-access-qwnf4\") pod \"packageserver-d55dfcdfc-slqwx\" (UID: \"9902dff6-4484-44ba-af31-877293debb91\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965254 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a6f67978-6c47-4927-ae94-1fef4eee7b18-cert\") pod \"ingress-canary-rw942\" (UID: \"a6f67978-6c47-4927-ae94-1fef4eee7b18\") " pod="openshift-ingress-canary/ingress-canary-rw942" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965293 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/cfccf8af-20ed-4d82-96d7-b4b3c00cd24b-mountpoint-dir\") pod \"csi-hostpathplugin-gczkp\" (UID: \"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b\") " pod="hostpath-provisioner/csi-hostpathplugin-gczkp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965314 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/cfccf8af-20ed-4d82-96d7-b4b3c00cd24b-registration-dir\") pod \"csi-hostpathplugin-gczkp\" (UID: \"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b\") " pod="hostpath-provisioner/csi-hostpathplugin-gczkp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965333 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/cfccf8af-20ed-4d82-96d7-b4b3c00cd24b-csi-data-dir\") pod \"csi-hostpathplugin-gczkp\" (UID: \"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b\") " pod="hostpath-provisioner/csi-hostpathplugin-gczkp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965352 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n26g7\" (UniqueName: \"kubernetes.io/projected/b3992349-bc61-405e-be75-609cce2fad10-kube-api-access-n26g7\") pod \"collect-profiles-29562300-vntp7\" (UID: \"b3992349-bc61-405e-be75-609cce2fad10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965373 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1507ff6e-1511-4c90-998e-1fb84da97e5b-proxy-tls\") pod \"machine-config-operator-74547568cd-vm474\" (UID: \"1507ff6e-1511-4c90-998e-1fb84da97e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vm474" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965397 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03c2477b-c56a-4354-bf4f-170b83f9240f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-2bpfb\" (UID: \"03c2477b-c56a-4354-bf4f-170b83f9240f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bpfb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965420 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9902dff6-4484-44ba-af31-877293debb91-apiservice-cert\") pod \"packageserver-d55dfcdfc-slqwx\" (UID: \"9902dff6-4484-44ba-af31-877293debb91\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965438 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f253587f-9f27-474d-8f46-21f1e6a79d97-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-lrhrs\" (UID: \"f253587f-9f27-474d-8f46-21f1e6a79d97\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lrhrs" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965466 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5x5b4\" (UniqueName: \"kubernetes.io/projected/f7c4987b-f8f6-4d39-b7f1-6f80c1384293-kube-api-access-5x5b4\") pod \"ingress-operator-5b745b69d9-s6zwp\" (UID: \"f7c4987b-f8f6-4d39-b7f1-6f80c1384293\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s6zwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965485 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z872t\" (UniqueName: \"kubernetes.io/projected/cfccf8af-20ed-4d82-96d7-b4b3c00cd24b-kube-api-access-z872t\") pod \"csi-hostpathplugin-gczkp\" (UID: \"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b\") " pod="hostpath-provisioner/csi-hostpathplugin-gczkp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965502 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/000e63dc-8228-486c-a6ee-12843810b5af-signing-cabundle\") pod \"service-ca-9c57cc56f-wttgv\" (UID: \"000e63dc-8228-486c-a6ee-12843810b5af\") " pod="openshift-service-ca/service-ca-9c57cc56f-wttgv" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965529 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/da36625d-0c68-48b8-9608-c7dedd304064-metrics-tls\") pod \"dns-default-cjn8k\" (UID: \"da36625d-0c68-48b8-9608-c7dedd304064\") " pod="openshift-dns/dns-default-cjn8k" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965570 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1507ff6e-1511-4c90-998e-1fb84da97e5b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vm474\" (UID: \"1507ff6e-1511-4c90-998e-1fb84da97e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vm474" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965588 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/37f3773a-2367-411b-b8e4-c4ae96970de3-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-j5djq\" (UID: \"37f3773a-2367-411b-b8e4-c4ae96970de3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5djq" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965631 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3992349-bc61-405e-be75-609cce2fad10-secret-volume\") pod \"collect-profiles-29562300-vntp7\" (UID: \"b3992349-bc61-405e-be75-609cce2fad10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965667 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq869\" (UniqueName: \"kubernetes.io/projected/173f4a11-926b-4cc0-a70c-c369ac37bac2-kube-api-access-pq869\") pod \"multus-admission-controller-857f4d67dd-6lsz5\" (UID: \"173f4a11-926b-4cc0-a70c-c369ac37bac2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6lsz5" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965687 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/03c2477b-c56a-4354-bf4f-170b83f9240f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-2bpfb\" (UID: \"03c2477b-c56a-4354-bf4f-170b83f9240f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bpfb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965702 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/000e63dc-8228-486c-a6ee-12843810b5af-signing-key\") pod \"service-ca-9c57cc56f-wttgv\" (UID: \"000e63dc-8228-486c-a6ee-12843810b5af\") " pod="openshift-service-ca/service-ca-9c57cc56f-wttgv" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965720 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f7c4987b-f8f6-4d39-b7f1-6f80c1384293-trusted-ca\") pod \"ingress-operator-5b745b69d9-s6zwp\" (UID: \"f7c4987b-f8f6-4d39-b7f1-6f80c1384293\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s6zwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965748 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f760de2f-292d-48fe-ac0c-a97aa37f556a-profile-collector-cert\") pod \"catalog-operator-68c6474976-8qg9n\" (UID: \"f760de2f-292d-48fe-ac0c-a97aa37f556a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8qg9n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965780 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/ed67dcc9-2a14-4e28-86df-da77a8d20012-certs\") pod \"machine-config-server-8hxbx\" (UID: \"ed67dcc9-2a14-4e28-86df-da77a8d20012\") " pod="openshift-machine-config-operator/machine-config-server-8hxbx" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965796 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ds7cb\" (UniqueName: \"kubernetes.io/projected/3abdbcf2-c3d0-4049-b08b-45fa02672dd4-kube-api-access-ds7cb\") pod \"package-server-manager-789f6589d5-7w9l5\" (UID: \"3abdbcf2-c3d0-4049-b08b-45fa02672dd4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7w9l5" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965821 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1507ff6e-1511-4c90-998e-1fb84da97e5b-images\") pod \"machine-config-operator-74547568cd-vm474\" (UID: \"1507ff6e-1511-4c90-998e-1fb84da97e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vm474" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965846 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f253587f-9f27-474d-8f46-21f1e6a79d97-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-lrhrs\" (UID: \"f253587f-9f27-474d-8f46-21f1e6a79d97\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lrhrs" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965873 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da36625d-0c68-48b8-9608-c7dedd304064-config-volume\") pod \"dns-default-cjn8k\" (UID: \"da36625d-0c68-48b8-9608-c7dedd304064\") " pod="openshift-dns/dns-default-cjn8k" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965889 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gssm9\" (UniqueName: \"kubernetes.io/projected/f253587f-9f27-474d-8f46-21f1e6a79d97-kube-api-access-gssm9\") pod \"kube-storage-version-migrator-operator-b67b599dd-lrhrs\" (UID: \"f253587f-9f27-474d-8f46-21f1e6a79d97\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lrhrs" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965916 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwn8d\" (UniqueName: \"kubernetes.io/projected/37f3773a-2367-411b-b8e4-c4ae96970de3-kube-api-access-cwn8d\") pod \"control-plane-machine-set-operator-78cbb6b69f-j5djq\" (UID: \"37f3773a-2367-411b-b8e4-c4ae96970de3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5djq" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965938 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/ed67dcc9-2a14-4e28-86df-da77a8d20012-node-bootstrap-token\") pod \"machine-config-server-8hxbx\" (UID: \"ed67dcc9-2a14-4e28-86df-da77a8d20012\") " pod="openshift-machine-config-operator/machine-config-server-8hxbx" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965954 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7f82q\" (UniqueName: \"kubernetes.io/projected/0021cb4b-86a2-46a5-83f6-109fee3e517b-kube-api-access-7f82q\") pod \"olm-operator-6b444d44fb-tjvgc\" (UID: \"0021cb4b-86a2-46a5-83f6-109fee3e517b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tjvgc" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965969 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9902dff6-4484-44ba-af31-877293debb91-tmpfs\") pod \"packageserver-d55dfcdfc-slqwx\" (UID: \"9902dff6-4484-44ba-af31-877293debb91\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.965986 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g44ww\" (UniqueName: \"kubernetes.io/projected/ed67dcc9-2a14-4e28-86df-da77a8d20012-kube-api-access-g44ww\") pod \"machine-config-server-8hxbx\" (UID: \"ed67dcc9-2a14-4e28-86df-da77a8d20012\") " pod="openshift-machine-config-operator/machine-config-server-8hxbx" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.966000 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3992349-bc61-405e-be75-609cce2fad10-config-volume\") pod \"collect-profiles-29562300-vntp7\" (UID: \"b3992349-bc61-405e-be75-609cce2fad10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.966019 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wslp6\" (UniqueName: \"kubernetes.io/projected/7954ca65-b7a4-47de-bc7c-2f57f0fee19a-kube-api-access-wslp6\") pod \"migrator-59844c95c7-tbl5t\" (UID: \"7954ca65-b7a4-47de-bc7c-2f57f0fee19a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tbl5t" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.966038 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/173f4a11-926b-4cc0-a70c-c369ac37bac2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6lsz5\" (UID: \"173f4a11-926b-4cc0-a70c-c369ac37bac2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6lsz5" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.966065 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d89e7afc-8f04-4996-b087-306543ef5acb-config\") pod \"service-ca-operator-777779d784-ghwhk\" (UID: \"d89e7afc-8f04-4996-b087-306543ef5acb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ghwhk" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.966087 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qr87f\" (UniqueName: \"kubernetes.io/projected/d89e7afc-8f04-4996-b087-306543ef5acb-kube-api-access-qr87f\") pod \"service-ca-operator-777779d784-ghwhk\" (UID: \"d89e7afc-8f04-4996-b087-306543ef5acb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ghwhk" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.966117 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcghd\" (UniqueName: \"kubernetes.io/projected/8626af43-6b5e-49e1-b19f-3986cdecdd2d-kube-api-access-fcghd\") pod \"auto-csr-approver-29562312-hjpcp\" (UID: \"8626af43-6b5e-49e1-b19f-3986cdecdd2d\") " pod="openshift-infra/auto-csr-approver-29562312-hjpcp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.966133 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/cfccf8af-20ed-4d82-96d7-b4b3c00cd24b-socket-dir\") pod \"csi-hostpathplugin-gczkp\" (UID: \"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b\") " pod="hostpath-provisioner/csi-hostpathplugin-gczkp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.966150 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6564\" (UniqueName: \"kubernetes.io/projected/da36625d-0c68-48b8-9608-c7dedd304064-kube-api-access-h6564\") pod \"dns-default-cjn8k\" (UID: \"da36625d-0c68-48b8-9608-c7dedd304064\") " pod="openshift-dns/dns-default-cjn8k" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.966166 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0021cb4b-86a2-46a5-83f6-109fee3e517b-srv-cert\") pod \"olm-operator-6b444d44fb-tjvgc\" (UID: \"0021cb4b-86a2-46a5-83f6-109fee3e517b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tjvgc" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.966183 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kpqgd\" (UniqueName: \"kubernetes.io/projected/a6f67978-6c47-4927-ae94-1fef4eee7b18-kube-api-access-kpqgd\") pod \"ingress-canary-rw942\" (UID: \"a6f67978-6c47-4927-ae94-1fef4eee7b18\") " pod="openshift-ingress-canary/ingress-canary-rw942" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.966199 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/58e4f2e1-b751-41e1-a56e-b62fd82b7caf-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-7tdh4\" (UID: \"58e4f2e1-b751-41e1-a56e-b62fd82b7caf\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7tdh4" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.966225 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9902dff6-4484-44ba-af31-877293debb91-webhook-cert\") pod \"packageserver-d55dfcdfc-slqwx\" (UID: \"9902dff6-4484-44ba-af31-877293debb91\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.966244 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/58e4f2e1-b751-41e1-a56e-b62fd82b7caf-proxy-tls\") pod \"machine-config-controller-84d6567774-7tdh4\" (UID: \"58e4f2e1-b751-41e1-a56e-b62fd82b7caf\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7tdh4" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.966267 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vg88g\" (UniqueName: \"kubernetes.io/projected/f760de2f-292d-48fe-ac0c-a97aa37f556a-kube-api-access-vg88g\") pod \"catalog-operator-68c6474976-8qg9n\" (UID: \"f760de2f-292d-48fe-ac0c-a97aa37f556a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8qg9n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.966295 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74f92156-f803-433a-99a4-981202076e81-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5wl9g\" (UID: \"74f92156-f803-433a-99a4-981202076e81\") " pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.966311 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0021cb4b-86a2-46a5-83f6-109fee3e517b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-tjvgc\" (UID: \"0021cb4b-86a2-46a5-83f6-109fee3e517b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tjvgc" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.967505 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1507ff6e-1511-4c90-998e-1fb84da97e5b-images\") pod \"machine-config-operator-74547568cd-vm474\" (UID: \"1507ff6e-1511-4c90-998e-1fb84da97e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vm474" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.968773 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/000e63dc-8228-486c-a6ee-12843810b5af-signing-cabundle\") pod \"service-ca-9c57cc56f-wttgv\" (UID: \"000e63dc-8228-486c-a6ee-12843810b5af\") " pod="openshift-service-ca/service-ca-9c57cc56f-wttgv" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.968810 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/cfccf8af-20ed-4d82-96d7-b4b3c00cd24b-mountpoint-dir\") pod \"csi-hostpathplugin-gczkp\" (UID: \"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b\") " pod="hostpath-provisioner/csi-hostpathplugin-gczkp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.968840 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/cfccf8af-20ed-4d82-96d7-b4b3c00cd24b-registration-dir\") pod \"csi-hostpathplugin-gczkp\" (UID: \"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b\") " pod="hostpath-provisioner/csi-hostpathplugin-gczkp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.968879 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/cfccf8af-20ed-4d82-96d7-b4b3c00cd24b-csi-data-dir\") pod \"csi-hostpathplugin-gczkp\" (UID: \"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b\") " pod="hostpath-provisioner/csi-hostpathplugin-gczkp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.968952 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0021cb4b-86a2-46a5-83f6-109fee3e517b-profile-collector-cert\") pod \"olm-operator-6b444d44fb-tjvgc\" (UID: \"0021cb4b-86a2-46a5-83f6-109fee3e517b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tjvgc" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.970286 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d89e7afc-8f04-4996-b087-306543ef5acb-config\") pod \"service-ca-operator-777779d784-ghwhk\" (UID: \"d89e7afc-8f04-4996-b087-306543ef5acb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ghwhk" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.970622 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/74f92156-f803-433a-99a4-981202076e81-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5wl9g\" (UID: \"74f92156-f803-433a-99a4-981202076e81\") " pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.971160 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1507ff6e-1511-4c90-998e-1fb84da97e5b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-vm474\" (UID: \"1507ff6e-1511-4c90-998e-1fb84da97e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vm474" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.971488 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/da36625d-0c68-48b8-9608-c7dedd304064-config-volume\") pod \"dns-default-cjn8k\" (UID: \"da36625d-0c68-48b8-9608-c7dedd304064\") " pod="openshift-dns/dns-default-cjn8k" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.971720 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/f7c4987b-f8f6-4d39-b7f1-6f80c1384293-metrics-tls\") pod \"ingress-operator-5b745b69d9-s6zwp\" (UID: \"f7c4987b-f8f6-4d39-b7f1-6f80c1384293\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s6zwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.971853 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/a6f67978-6c47-4927-ae94-1fef4eee7b18-cert\") pod \"ingress-canary-rw942\" (UID: \"a6f67978-6c47-4927-ae94-1fef4eee7b18\") " pod="openshift-ingress-canary/ingress-canary-rw942" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.972281 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/9902dff6-4484-44ba-af31-877293debb91-tmpfs\") pod \"packageserver-d55dfcdfc-slqwx\" (UID: \"9902dff6-4484-44ba-af31-877293debb91\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.972316 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1507ff6e-1511-4c90-998e-1fb84da97e5b-proxy-tls\") pod \"machine-config-operator-74547568cd-vm474\" (UID: \"1507ff6e-1511-4c90-998e-1fb84da97e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vm474" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.972326 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3abdbcf2-c3d0-4049-b08b-45fa02672dd4-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-7w9l5\" (UID: \"3abdbcf2-c3d0-4049-b08b-45fa02672dd4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7w9l5" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.972769 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/da36625d-0c68-48b8-9608-c7dedd304064-metrics-tls\") pod \"dns-default-cjn8k\" (UID: \"da36625d-0c68-48b8-9608-c7dedd304064\") " pod="openshift-dns/dns-default-cjn8k" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.972868 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f760de2f-292d-48fe-ac0c-a97aa37f556a-srv-cert\") pod \"catalog-operator-68c6474976-8qg9n\" (UID: \"f760de2f-292d-48fe-ac0c-a97aa37f556a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8qg9n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.972871 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d89e7afc-8f04-4996-b087-306543ef5acb-serving-cert\") pod \"service-ca-operator-777779d784-ghwhk\" (UID: \"d89e7afc-8f04-4996-b087-306543ef5acb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ghwhk" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.975071 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f253587f-9f27-474d-8f46-21f1e6a79d97-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-lrhrs\" (UID: \"f253587f-9f27-474d-8f46-21f1e6a79d97\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lrhrs" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.975212 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f253587f-9f27-474d-8f46-21f1e6a79d97-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-lrhrs\" (UID: \"f253587f-9f27-474d-8f46-21f1e6a79d97\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lrhrs" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.975263 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/cfccf8af-20ed-4d82-96d7-b4b3c00cd24b-socket-dir\") pod \"csi-hostpathplugin-gczkp\" (UID: \"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b\") " pod="hostpath-provisioner/csi-hostpathplugin-gczkp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.976086 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3992349-bc61-405e-be75-609cce2fad10-config-volume\") pod \"collect-profiles-29562300-vntp7\" (UID: \"b3992349-bc61-405e-be75-609cce2fad10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.976112 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/58e4f2e1-b751-41e1-a56e-b62fd82b7caf-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-7tdh4\" (UID: \"58e4f2e1-b751-41e1-a56e-b62fd82b7caf\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7tdh4" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.976334 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/58e4f2e1-b751-41e1-a56e-b62fd82b7caf-proxy-tls\") pod \"machine-config-controller-84d6567774-7tdh4\" (UID: \"58e4f2e1-b751-41e1-a56e-b62fd82b7caf\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7tdh4" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.976370 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/ed67dcc9-2a14-4e28-86df-da77a8d20012-node-bootstrap-token\") pod \"machine-config-server-8hxbx\" (UID: \"ed67dcc9-2a14-4e28-86df-da77a8d20012\") " pod="openshift-machine-config-operator/machine-config-server-8hxbx" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.978732 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/37f3773a-2367-411b-b8e4-c4ae96970de3-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-j5djq\" (UID: \"37f3773a-2367-411b-b8e4-c4ae96970de3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5djq" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.978856 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74f92156-f803-433a-99a4-981202076e81-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5wl9g\" (UID: \"74f92156-f803-433a-99a4-981202076e81\") " pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.978962 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/9902dff6-4484-44ba-af31-877293debb91-webhook-cert\") pod \"packageserver-d55dfcdfc-slqwx\" (UID: \"9902dff6-4484-44ba-af31-877293debb91\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.978992 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f760de2f-292d-48fe-ac0c-a97aa37f556a-profile-collector-cert\") pod \"catalog-operator-68c6474976-8qg9n\" (UID: \"f760de2f-292d-48fe-ac0c-a97aa37f556a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8qg9n" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.979405 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/ed67dcc9-2a14-4e28-86df-da77a8d20012-certs\") pod \"machine-config-server-8hxbx\" (UID: \"ed67dcc9-2a14-4e28-86df-da77a8d20012\") " pod="openshift-machine-config-operator/machine-config-server-8hxbx" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.979608 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/9902dff6-4484-44ba-af31-877293debb91-apiservice-cert\") pod \"packageserver-d55dfcdfc-slqwx\" (UID: \"9902dff6-4484-44ba-af31-877293debb91\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.979699 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/03c2477b-c56a-4354-bf4f-170b83f9240f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-2bpfb\" (UID: \"03c2477b-c56a-4354-bf4f-170b83f9240f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bpfb" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.979962 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/f7c4987b-f8f6-4d39-b7f1-6f80c1384293-trusted-ca\") pod \"ingress-operator-5b745b69d9-s6zwp\" (UID: \"f7c4987b-f8f6-4d39-b7f1-6f80c1384293\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s6zwp" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.981270 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0021cb4b-86a2-46a5-83f6-109fee3e517b-srv-cert\") pod \"olm-operator-6b444d44fb-tjvgc\" (UID: \"0021cb4b-86a2-46a5-83f6-109fee3e517b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tjvgc" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.981277 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3992349-bc61-405e-be75-609cce2fad10-secret-volume\") pod \"collect-profiles-29562300-vntp7\" (UID: \"b3992349-bc61-405e-be75-609cce2fad10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.982323 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2bd4438d-8a23-4e8b-aad7-f5775f1d88a8-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-wwbwr\" (UID: \"2bd4438d-8a23-4e8b-aad7-f5775f1d88a8\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwbwr" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.982485 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4rqzj" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.983315 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/000e63dc-8228-486c-a6ee-12843810b5af-signing-key\") pod \"service-ca-9c57cc56f-wttgv\" (UID: \"000e63dc-8228-486c-a6ee-12843810b5af\") " pod="openshift-service-ca/service-ca-9c57cc56f-wttgv" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.985875 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/173f4a11-926b-4cc0-a70c-c369ac37bac2-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-6lsz5\" (UID: \"173f4a11-926b-4cc0-a70c-c369ac37bac2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6lsz5" Mar 17 09:13:43 crc kubenswrapper[4813]: I0317 09:13:43.996845 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8745x\" (UniqueName: \"kubernetes.io/projected/93cb7b2d-1c40-40be-82d7-1858272444ec-kube-api-access-8745x\") pod \"cluster-image-registry-operator-dc59b4c8b-njr98\" (UID: \"93cb7b2d-1c40-40be-82d7-1858272444ec\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-njr98" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.017795 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2657\" (UniqueName: \"kubernetes.io/projected/550c18c5-f4e0-4da1-9692-c6cacfa882af-kube-api-access-z2657\") pod \"authentication-operator-69f744f599-tzt6n\" (UID: \"550c18c5-f4e0-4da1-9692-c6cacfa882af\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-tzt6n" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.036405 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d32fb60f-2d76-4e02-9839-535e1543eff7-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-s2hhg\" (UID: \"d32fb60f-2d76-4e02-9839-535e1543eff7\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s2hhg" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.049690 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l"] Mar 17 09:13:44 crc kubenswrapper[4813]: W0317 09:13:44.059420 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod768d1a31_0d78_4239_9d7d_260d11e4ad58.slice/crio-6a5b92ae48f0bc497af13d4855416ea5532d65c8ef4a5339edb0bb79b8b84d4b WatchSource:0}: Error finding container 6a5b92ae48f0bc497af13d4855416ea5532d65c8ef4a5339edb0bb79b8b84d4b: Status 404 returned error can't find the container with id 6a5b92ae48f0bc497af13d4855416ea5532d65c8ef4a5339edb0bb79b8b84d4b Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.060107 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4lhgk\" (UniqueName: \"kubernetes.io/projected/68399978-6f43-4458-a1e1-58d5984de8f5-kube-api-access-4lhgk\") pod \"router-default-5444994796-m97th\" (UID: \"68399978-6f43-4458-a1e1-58d5984de8f5\") " pod="openshift-ingress/router-default-5444994796-m97th" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.072473 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:44 crc kubenswrapper[4813]: E0317 09:13:44.074302 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:44.574285098 +0000 UTC m=+246.675088597 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.080647 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s29cv\" (UniqueName: \"kubernetes.io/projected/12efeac2-9fc3-425a-8874-3c2d957e04ca-kube-api-access-s29cv\") pod \"machine-approver-56656f9798-kpxlg\" (UID: \"12efeac2-9fc3-425a-8874-3c2d957e04ca\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kpxlg" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.081662 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-m97th" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.096920 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-njr98" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.101003 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f2q6\" (UniqueName: \"kubernetes.io/projected/94f282fa-7b22-41e3-9911-073b05d41200-kube-api-access-7f2q6\") pod \"dns-operator-744455d44c-qwlbl\" (UID: \"94f282fa-7b22-41e3-9911-073b05d41200\") " pod="openshift-dns-operator/dns-operator-744455d44c-qwlbl" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.103232 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s2hhg" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.108157 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.115539 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.115616 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.122425 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24qg4\" (UniqueName: \"kubernetes.io/projected/3aa0fc32-ed11-482b-9a6d-fcd59544bf8d-kube-api-access-24qg4\") pod \"downloads-7954f5f757-xb4k9\" (UID: \"3aa0fc32-ed11-482b-9a6d-fcd59544bf8d\") " pod="openshift-console/downloads-7954f5f757-xb4k9" Mar 17 09:13:44 crc kubenswrapper[4813]: W0317 09:13:44.131566 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68399978_6f43_4458_a1e1_58d5984de8f5.slice/crio-43fb4a88403ac82b2d4342a177665b23b725601b7987976c4df2d71e053634ef WatchSource:0}: Error finding container 43fb4a88403ac82b2d4342a177665b23b725601b7987976c4df2d71e053634ef: Status 404 returned error can't find the container with id 43fb4a88403ac82b2d4342a177665b23b725601b7987976c4df2d71e053634ef Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.131674 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwbwr" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.138473 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cnpcn\" (UniqueName: \"kubernetes.io/projected/845ee13b-7ed9-4910-958b-5269812237f0-kube-api-access-cnpcn\") pod \"cluster-samples-operator-665b6dd947-dpghh\" (UID: \"845ee13b-7ed9-4910-958b-5269812237f0\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dpghh" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.160788 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dx7gs\" (UniqueName: \"kubernetes.io/projected/364cd6ac-033f-455e-a95a-10d50f055fb2-kube-api-access-dx7gs\") pod \"etcd-operator-b45778765-jzbwp\" (UID: \"364cd6ac-033f-455e-a95a-10d50f055fb2\") " pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.174279 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:44 crc kubenswrapper[4813]: E0317 09:13:44.174638 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:44.674620804 +0000 UTC m=+246.775424303 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.191862 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngkkm\" (UniqueName: \"kubernetes.io/projected/f7603094-bb1d-4d77-b463-2270003b2805-kube-api-access-ngkkm\") pod \"console-f9d7485db-2w9vb\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.202148 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-4rqzj"] Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.205105 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzrsb\" (UniqueName: \"kubernetes.io/projected/72fa1c68-872b-45a3-98b7-b0b1ebc08ad7-kube-api-access-tzrsb\") pod \"openshift-controller-manager-operator-756b6f6bc6-wcd9z\" (UID: \"72fa1c68-872b-45a3-98b7-b0b1ebc08ad7\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wcd9z" Mar 17 09:13:44 crc kubenswrapper[4813]: W0317 09:13:44.218665 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb828a0e2_3def_4bb5_8f54_2792376423cd.slice/crio-d84acf03330d40fc53f0b38991c3920a7c60d478f1275eeac36689d4e6d6a9a3 WatchSource:0}: Error finding container d84acf03330d40fc53f0b38991c3920a7c60d478f1275eeac36689d4e6d6a9a3: Status 404 returned error can't find the container with id d84acf03330d40fc53f0b38991c3920a7c60d478f1275eeac36689d4e6d6a9a3 Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.245980 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-tzt6n" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.246521 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8hk6\" (UniqueName: \"kubernetes.io/projected/74f92156-f803-433a-99a4-981202076e81-kube-api-access-q8hk6\") pod \"marketplace-operator-79b997595-5wl9g\" (UID: \"74f92156-f803-433a-99a4-981202076e81\") " pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.262141 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zgwpz\" (UniqueName: \"kubernetes.io/projected/000e63dc-8228-486c-a6ee-12843810b5af-kube-api-access-zgwpz\") pod \"service-ca-9c57cc56f-wttgv\" (UID: \"000e63dc-8228-486c-a6ee-12843810b5af\") " pod="openshift-service-ca/service-ca-9c57cc56f-wttgv" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.273297 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kpxlg" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.275388 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:44 crc kubenswrapper[4813]: E0317 09:13:44.275562 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:44.775539871 +0000 UTC m=+246.876343370 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.275652 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.276089 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nk8pf\" (UniqueName: \"kubernetes.io/projected/1507ff6e-1511-4c90-998e-1fb84da97e5b-kube-api-access-nk8pf\") pod \"machine-config-operator-74547568cd-vm474\" (UID: \"1507ff6e-1511-4c90-998e-1fb84da97e5b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vm474" Mar 17 09:13:44 crc kubenswrapper[4813]: E0317 09:13:44.276769 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:44.776752393 +0000 UTC m=+246.877555892 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.292422 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dpghh" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.311641 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/f7c4987b-f8f6-4d39-b7f1-6f80c1384293-bound-sa-token\") pod \"ingress-operator-5b745b69d9-s6zwp\" (UID: \"f7c4987b-f8f6-4d39-b7f1-6f80c1384293\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s6zwp" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.317620 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlzt2\" (UniqueName: \"kubernetes.io/projected/58e4f2e1-b751-41e1-a56e-b62fd82b7caf-kube-api-access-tlzt2\") pod \"machine-config-controller-84d6567774-7tdh4\" (UID: \"58e4f2e1-b751-41e1-a56e-b62fd82b7caf\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7tdh4" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.319880 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wcd9z" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.343107 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwnf4\" (UniqueName: \"kubernetes.io/projected/9902dff6-4484-44ba-af31-877293debb91-kube-api-access-qwnf4\") pod \"packageserver-d55dfcdfc-slqwx\" (UID: \"9902dff6-4484-44ba-af31-877293debb91\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.362407 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-xb4k9" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.366639 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwbwr"] Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.366294 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n26g7\" (UniqueName: \"kubernetes.io/projected/b3992349-bc61-405e-be75-609cce2fad10-kube-api-access-n26g7\") pod \"collect-profiles-29562300-vntp7\" (UID: \"b3992349-bc61-405e-be75-609cce2fad10\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.380495 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:44 crc kubenswrapper[4813]: E0317 09:13:44.380614 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:44.880581489 +0000 UTC m=+246.981384988 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.380722 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:44 crc kubenswrapper[4813]: E0317 09:13:44.381064 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:44.881057135 +0000 UTC m=+246.981860634 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.389558 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-qwlbl" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.390343 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/03c2477b-c56a-4354-bf4f-170b83f9240f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-2bpfb\" (UID: \"03c2477b-c56a-4354-bf4f-170b83f9240f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bpfb" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.406298 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s2hhg"] Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.408384 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gssm9\" (UniqueName: \"kubernetes.io/projected/f253587f-9f27-474d-8f46-21f1e6a79d97-kube-api-access-gssm9\") pod \"kube-storage-version-migrator-operator-b67b599dd-lrhrs\" (UID: \"f253587f-9f27-474d-8f46-21f1e6a79d97\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lrhrs" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.409195 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.415946 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.419171 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwn8d\" (UniqueName: \"kubernetes.io/projected/37f3773a-2367-411b-b8e4-c4ae96970de3-kube-api-access-cwn8d\") pod \"control-plane-machine-set-operator-78cbb6b69f-j5djq\" (UID: \"37f3773a-2367-411b-b8e4-c4ae96970de3\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5djq" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.439470 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g44ww\" (UniqueName: \"kubernetes.io/projected/ed67dcc9-2a14-4e28-86df-da77a8d20012-kube-api-access-g44ww\") pod \"machine-config-server-8hxbx\" (UID: \"ed67dcc9-2a14-4e28-86df-da77a8d20012\") " pod="openshift-machine-config-operator/machine-config-server-8hxbx" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.450853 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5djq" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.453617 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lrhrs" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.459441 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-njr98"] Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.459891 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vm474" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.461944 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7f82q\" (UniqueName: \"kubernetes.io/projected/0021cb4b-86a2-46a5-83f6-109fee3e517b-kube-api-access-7f82q\") pod \"olm-operator-6b444d44fb-tjvgc\" (UID: \"0021cb4b-86a2-46a5-83f6-109fee3e517b\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tjvgc" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.475502 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.480189 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ds7cb\" (UniqueName: \"kubernetes.io/projected/3abdbcf2-c3d0-4049-b08b-45fa02672dd4-kube-api-access-ds7cb\") pod \"package-server-manager-789f6589d5-7w9l5\" (UID: \"3abdbcf2-c3d0-4049-b08b-45fa02672dd4\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7w9l5" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.481553 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.481932 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bpfb" Mar 17 09:13:44 crc kubenswrapper[4813]: E0317 09:13:44.482041 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:44.982014902 +0000 UTC m=+247.082818401 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.482423 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:44 crc kubenswrapper[4813]: E0317 09:13:44.482934 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:44.982924974 +0000 UTC m=+247.083728473 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.501377 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5x5b4\" (UniqueName: \"kubernetes.io/projected/f7c4987b-f8f6-4d39-b7f1-6f80c1384293-kube-api-access-5x5b4\") pod \"ingress-operator-5b745b69d9-s6zwp\" (UID: \"f7c4987b-f8f6-4d39-b7f1-6f80c1384293\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s6zwp" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.518483 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s6zwp" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.523664 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z872t\" (UniqueName: \"kubernetes.io/projected/cfccf8af-20ed-4d82-96d7-b4b3c00cd24b-kube-api-access-z872t\") pod \"csi-hostpathplugin-gczkp\" (UID: \"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b\") " pod="hostpath-provisioner/csi-hostpathplugin-gczkp" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.525296 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-wttgv" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.563767 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.564465 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7w9l5" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.565733 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7tdh4" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.571534 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.571912 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tjvgc" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.572371 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vg88g\" (UniqueName: \"kubernetes.io/projected/f760de2f-292d-48fe-ac0c-a97aa37f556a-kube-api-access-vg88g\") pod \"catalog-operator-68c6474976-8qg9n\" (UID: \"f760de2f-292d-48fe-ac0c-a97aa37f556a\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8qg9n" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.582427 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-8hxbx" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.583638 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:44 crc kubenswrapper[4813]: E0317 09:13:44.584112 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:45.084093659 +0000 UTC m=+247.184897148 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.594499 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qr87f\" (UniqueName: \"kubernetes.io/projected/d89e7afc-8f04-4996-b087-306543ef5acb-kube-api-access-qr87f\") pod \"service-ca-operator-777779d784-ghwhk\" (UID: \"d89e7afc-8f04-4996-b087-306543ef5acb\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-ghwhk" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.595955 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kpqgd\" (UniqueName: \"kubernetes.io/projected/a6f67978-6c47-4927-ae94-1fef4eee7b18-kube-api-access-kpqgd\") pod \"ingress-canary-rw942\" (UID: \"a6f67978-6c47-4927-ae94-1fef4eee7b18\") " pod="openshift-ingress-canary/ingress-canary-rw942" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.609228 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcghd\" (UniqueName: \"kubernetes.io/projected/8626af43-6b5e-49e1-b19f-3986cdecdd2d-kube-api-access-fcghd\") pod \"auto-csr-approver-29562312-hjpcp\" (UID: \"8626af43-6b5e-49e1-b19f-3986cdecdd2d\") " pod="openshift-infra/auto-csr-approver-29562312-hjpcp" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.609427 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-tzt6n"] Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.610243 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dpghh"] Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.634290 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wcd9z"] Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.639106 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wslp6\" (UniqueName: \"kubernetes.io/projected/7954ca65-b7a4-47de-bc7c-2f57f0fee19a-kube-api-access-wslp6\") pod \"migrator-59844c95c7-tbl5t\" (UID: \"7954ca65-b7a4-47de-bc7c-2f57f0fee19a\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tbl5t" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.648174 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6564\" (UniqueName: \"kubernetes.io/projected/da36625d-0c68-48b8-9608-c7dedd304064-kube-api-access-h6564\") pod \"dns-default-cjn8k\" (UID: \"da36625d-0c68-48b8-9608-c7dedd304064\") " pod="openshift-dns/dns-default-cjn8k" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.659853 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq869\" (UniqueName: \"kubernetes.io/projected/173f4a11-926b-4cc0-a70c-c369ac37bac2-kube-api-access-pq869\") pod \"multus-admission-controller-857f4d67dd-6lsz5\" (UID: \"173f4a11-926b-4cc0-a70c-c369ac37bac2\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-6lsz5" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.674356 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-gczkp" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.685284 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:44 crc kubenswrapper[4813]: E0317 09:13:44.685589 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:45.185577316 +0000 UTC m=+247.286380815 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:44 crc kubenswrapper[4813]: W0317 09:13:44.712076 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod72fa1c68_872b_45a3_98b7_b0b1ebc08ad7.slice/crio-5152061a4c8e824ece65b39efb5577ea7a720643573e20f00c864e375d85d3cf WatchSource:0}: Error finding container 5152061a4c8e824ece65b39efb5577ea7a720643573e20f00c864e375d85d3cf: Status 404 returned error can't find the container with id 5152061a4c8e824ece65b39efb5577ea7a720643573e20f00c864e375d85d3cf Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.737468 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tbl5t" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.770476 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-6lsz5" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.786323 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:44 crc kubenswrapper[4813]: E0317 09:13:44.786482 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:45.286457191 +0000 UTC m=+247.387260690 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.786954 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:44 crc kubenswrapper[4813]: E0317 09:13:44.787226 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:45.287218616 +0000 UTC m=+247.388022105 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.788227 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8qg9n" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.813220 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562312-hjpcp" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.835774 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ghwhk" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.887500 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-cjn8k" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.890266 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:44 crc kubenswrapper[4813]: E0317 09:13:44.890754 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:45.390735632 +0000 UTC m=+247.491539131 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.892560 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-rw942" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.907915 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-qwlbl"] Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.910807 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-jzbwp"] Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.935727 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-xb4k9"] Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.938811 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kpxlg" event={"ID":"12efeac2-9fc3-425a-8874-3c2d957e04ca","Type":"ContainerStarted","Data":"b019b7ffcc3acf876976fd754fcd67edb37b8090be310807630ed91ad921d110"} Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.940903 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-m97th" event={"ID":"68399978-6f43-4458-a1e1-58d5984de8f5","Type":"ContainerStarted","Data":"5c19b93e8fb645faf7a061438c2192463f2628452377a2a286e50f4fd5f2fc29"} Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.940977 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-m97th" event={"ID":"68399978-6f43-4458-a1e1-58d5984de8f5","Type":"ContainerStarted","Data":"43fb4a88403ac82b2d4342a177665b23b725601b7987976c4df2d71e053634ef"} Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.942077 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wcd9z" event={"ID":"72fa1c68-872b-45a3-98b7-b0b1ebc08ad7","Type":"ContainerStarted","Data":"5152061a4c8e824ece65b39efb5577ea7a720643573e20f00c864e375d85d3cf"} Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.943874 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kns96" event={"ID":"26a86c9b-1645-49c2-a2fb-78b210984363","Type":"ContainerStarted","Data":"6856613ffaaecea4d97cff934d9c8369fab91f9ba8ff34b3e828c5f841c24f99"} Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.943903 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-kns96" event={"ID":"26a86c9b-1645-49c2-a2fb-78b210984363","Type":"ContainerStarted","Data":"a0e12e048a946503f3f6e17e894ee74c3362dba8523f986475940e83f5aae80e"} Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.944980 4813 generic.go:334] "Generic (PLEG): container finished" podID="b828a0e2-3def-4bb5-8f54-2792376423cd" containerID="10c59952e2a4beef6ca9231c78c721102bc2fc3216a6830bb98551c0755ebf04" exitCode=0 Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.945051 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4rqzj" event={"ID":"b828a0e2-3def-4bb5-8f54-2792376423cd","Type":"ContainerDied","Data":"10c59952e2a4beef6ca9231c78c721102bc2fc3216a6830bb98551c0755ebf04"} Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.945075 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4rqzj" event={"ID":"b828a0e2-3def-4bb5-8f54-2792376423cd","Type":"ContainerStarted","Data":"d84acf03330d40fc53f0b38991c3920a7c60d478f1275eeac36689d4e6d6a9a3"} Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.966020 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-2fmh2" event={"ID":"09bf1deb-b09d-4978-baac-c06782c2ca2f","Type":"ContainerStarted","Data":"d0d975c121ffab031efe832451190254bb2ae9591c4fb88891b56a5e7d0b3315"} Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.966055 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-2fmh2" event={"ID":"09bf1deb-b09d-4978-baac-c06782c2ca2f","Type":"ContainerStarted","Data":"cec88584e36fa2c7b2c24eae6d339947066a99b763bfbc036be3d417b49379e3"} Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.966233 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-2fmh2" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.981115 4813 patch_prober.go:28] interesting pod/console-operator-58897d9998-2fmh2 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.981156 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-2fmh2" podUID="09bf1deb-b09d-4978-baac-c06782c2ca2f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.989550 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" event={"ID":"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0","Type":"ContainerStarted","Data":"a9484cecf8ed9d637c1ce56409a51a79ed0919e00abef61a98204d77038b3650"} Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.989613 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" event={"ID":"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0","Type":"ContainerStarted","Data":"0a9009f31556196a712dbea20111012eb1c927a32e77a31daaae2fc5e415da10"} Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.990047 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.991117 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:44 crc kubenswrapper[4813]: E0317 09:13:44.991419 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:45.491409171 +0000 UTC m=+247.592212670 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:44 crc kubenswrapper[4813]: I0317 09:13:44.994646 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-tzt6n" event={"ID":"550c18c5-f4e0-4da1-9692-c6cacfa882af","Type":"ContainerStarted","Data":"d3257dad4e2dd9c4610e85d0d653aac6f08274f142e611e6c81e07c032eca92c"} Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:44.998054 4813 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-5kp7w container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" start-of-body= Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:44.998101 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" podUID="5e837d92-c48e-4b0f-a55a-40bdaea5d9f0" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.11:6443/healthz\": dial tcp 10.217.0.11:6443: connect: connection refused" Mar 17 09:13:45 crc kubenswrapper[4813]: W0317 09:13:44.999458 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod94f282fa_7b22_41e3_9911_073b05d41200.slice/crio-9e83d12e0cb648d54fc684b8eebbb051c735c0472d6a0708b530175d04e0b3d5 WatchSource:0}: Error finding container 9e83d12e0cb648d54fc684b8eebbb051c735c0472d6a0708b530175d04e0b3d5: Status 404 returned error can't find the container with id 9e83d12e0cb648d54fc684b8eebbb051c735c0472d6a0708b530175d04e0b3d5 Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.001299 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-njr98" event={"ID":"93cb7b2d-1c40-40be-82d7-1858272444ec","Type":"ContainerStarted","Data":"238a6e5864761ebc4449a654a22022f964edd3d5f0952891699535d0e2306ef4"} Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.001671 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-wttgv"] Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.005694 4813 generic.go:334] "Generic (PLEG): container finished" podID="eb147d37-445a-48f2-a2e6-fb16f3575529" containerID="7acf945901f6e88018e5707cb3b2bedb706fa516349910c0610e9e5420f978f0" exitCode=0 Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.006109 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" event={"ID":"eb147d37-445a-48f2-a2e6-fb16f3575529","Type":"ContainerDied","Data":"7acf945901f6e88018e5707cb3b2bedb706fa516349910c0610e9e5420f978f0"} Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.006232 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" event={"ID":"eb147d37-445a-48f2-a2e6-fb16f3575529","Type":"ContainerStarted","Data":"95e77e3f1afe36af208b6ee8400904d1eb126d3f625a2467bd94cc39335dc94a"} Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.015857 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" event={"ID":"768d1a31-0d78-4239-9d7d-260d11e4ad58","Type":"ContainerStarted","Data":"120c9ea27cd7b3db8e036a33512fb311f5c7126c479216e881ff6ac734fdda55"} Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.015895 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" event={"ID":"768d1a31-0d78-4239-9d7d-260d11e4ad58","Type":"ContainerStarted","Data":"6a5b92ae48f0bc497af13d4855416ea5532d65c8ef4a5339edb0bb79b8b84d4b"} Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.016177 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.022952 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s2hhg" event={"ID":"d32fb60f-2d76-4e02-9839-535e1543eff7","Type":"ContainerStarted","Data":"b95408ae145ddc58546ac2787fe95fc83871550bbd9cb7313a6c4b3d10757d8a"} Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.028428 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k58b" event={"ID":"98211622-4794-4412-a9d0-64436ee1a3f4","Type":"ContainerStarted","Data":"7d67bb9435360b6790cd3ca4f9b4dc4635fb1293c141753c67ce1a1f37162c5d"} Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.028473 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k58b" event={"ID":"98211622-4794-4412-a9d0-64436ee1a3f4","Type":"ContainerStarted","Data":"505f761af0bd804e9abeacc49ac4f0c6abc151024c424e6073a665c371883f06"} Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.036698 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwbwr" event={"ID":"2bd4438d-8a23-4e8b-aad7-f5775f1d88a8","Type":"ContainerStarted","Data":"7107b434a486745184e0af67da283cfc2788036c7a3218f3fcf8fc2204d88463"} Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.040366 4813 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-6rn9l container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.040417 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" podUID="768d1a31-0d78-4239-9d7d-260d11e4ad58" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.8:8443/healthz\": dial tcp 10.217.0.8:8443: connect: connection refused" Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.041079 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" event={"ID":"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30","Type":"ContainerStarted","Data":"168060a45f4c11cf0e1cbc8021564701a55997c1a91b4a487bb37581d4c1f7d8"} Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.041523 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.056864 4813 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-ptq22 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.056930 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" podUID="55bb4fe2-94b0-4fdf-abb4-e39470fe2c30" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 17 09:13:45 crc kubenswrapper[4813]: W0317 09:13:45.078054 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3aa0fc32_ed11_482b_9a6d_fcd59544bf8d.slice/crio-5f2a3af66e994d421f9e46466a63fdd84eb30b09966a56680ee4d7ca97d76eea WatchSource:0}: Error finding container 5f2a3af66e994d421f9e46466a63fdd84eb30b09966a56680ee4d7ca97d76eea: Status 404 returned error can't find the container with id 5f2a3af66e994d421f9e46466a63fdd84eb30b09966a56680ee4d7ca97d76eea Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.082040 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-m97th" Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.085256 4813 patch_prober.go:28] interesting pod/router-default-5444994796-m97th container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.085332 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-m97th" podUID="68399978-6f43-4458-a1e1-58d5984de8f5" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.091530 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:45 crc kubenswrapper[4813]: E0317 09:13:45.091722 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:45.591694696 +0000 UTC m=+247.692498195 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.091859 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:45 crc kubenswrapper[4813]: E0317 09:13:45.095151 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:45.595135832 +0000 UTC m=+247.695939331 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.143897 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lrhrs"] Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.147475 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bpfb"] Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.152769 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5djq"] Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.161382 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-2w9vb"] Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.196489 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:45 crc kubenswrapper[4813]: E0317 09:13:45.196926 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:45.696884337 +0000 UTC m=+247.797687836 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.197297 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:45 crc kubenswrapper[4813]: E0317 09:13:45.216867 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:45.716851588 +0000 UTC m=+247.817655087 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.231536 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5wl9g"] Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.261635 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7w9l5"] Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.300512 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:45 crc kubenswrapper[4813]: E0317 09:13:45.300822 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:45.800809037 +0000 UTC m=+247.901612526 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.340336 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-7tdh4"] Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.363569 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-vm474"] Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.409319 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:45 crc kubenswrapper[4813]: E0317 09:13:45.413482 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:45.913466944 +0000 UTC m=+248.014270443 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:45 crc kubenswrapper[4813]: W0317 09:13:45.483577 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod74f92156_f803_433a_99a4_981202076e81.slice/crio-fca401c8c10481f6181cc4f34ed269a2838d6d313adba4fdfe6a786a51f43e33 WatchSource:0}: Error finding container fca401c8c10481f6181cc4f34ed269a2838d6d313adba4fdfe6a786a51f43e33: Status 404 returned error can't find the container with id fca401c8c10481f6181cc4f34ed269a2838d6d313adba4fdfe6a786a51f43e33 Mar 17 09:13:45 crc kubenswrapper[4813]: W0317 09:13:45.499854 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod58e4f2e1_b751_41e1_a56e_b62fd82b7caf.slice/crio-58d96a4f3b77cc079a04e120af2e3fe6b99ba21096f3f334981d5ed845852e30 WatchSource:0}: Error finding container 58d96a4f3b77cc079a04e120af2e3fe6b99ba21096f3f334981d5ed845852e30: Status 404 returned error can't find the container with id 58d96a4f3b77cc079a04e120af2e3fe6b99ba21096f3f334981d5ed845852e30 Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.512615 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:45 crc kubenswrapper[4813]: E0317 09:13:45.512900 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:46.012884569 +0000 UTC m=+248.113688068 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:45 crc kubenswrapper[4813]: W0317 09:13:45.540032 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1507ff6e_1511_4c90_998e_1fb84da97e5b.slice/crio-bac03690eff98f2bbc9a66c0639171baa801c676d8809043675f32ecad54c1d2 WatchSource:0}: Error finding container bac03690eff98f2bbc9a66c0639171baa801c676d8809043675f32ecad54c1d2: Status 404 returned error can't find the container with id bac03690eff98f2bbc9a66c0639171baa801c676d8809043675f32ecad54c1d2 Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.622452 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:45 crc kubenswrapper[4813]: E0317 09:13:45.622831 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:46.122816803 +0000 UTC m=+248.223620302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.627159 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx"] Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.723069 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:45 crc kubenswrapper[4813]: E0317 09:13:45.723457 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:46.223410328 +0000 UTC m=+248.324213827 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.824536 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:45 crc kubenswrapper[4813]: E0317 09:13:45.825040 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:46.325029599 +0000 UTC m=+248.425833098 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.879926 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-gczkp"] Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.929840 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-sk47b" podStartSLOduration=183.929822198 podStartE2EDuration="3m3.929822198s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:45.898320875 +0000 UTC m=+247.999124374" watchObservedRunningTime="2026-03-17 09:13:45.929822198 +0000 UTC m=+248.030625697" Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.930292 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:45 crc kubenswrapper[4813]: I0317 09:13:45.930686 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7"] Mar 17 09:13:45 crc kubenswrapper[4813]: E0317 09:13:45.930713 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:46.430698688 +0000 UTC m=+248.531502177 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.017080 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tjvgc"] Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.032522 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-tbl5t"] Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.033078 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:46 crc kubenswrapper[4813]: E0317 09:13:46.035493 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:46.535480036 +0000 UTC m=+248.636283525 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.078647 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vm474" event={"ID":"1507ff6e-1511-4c90-998e-1fb84da97e5b","Type":"ContainerStarted","Data":"bac03690eff98f2bbc9a66c0639171baa801c676d8809043675f32ecad54c1d2"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.088375 4813 patch_prober.go:28] interesting pod/router-default-5444994796-m97th container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 09:13:46 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Mar 17 09:13:46 crc kubenswrapper[4813]: [+]process-running ok Mar 17 09:13:46 crc kubenswrapper[4813]: healthz check failed Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.088432 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-m97th" podUID="68399978-6f43-4458-a1e1-58d5984de8f5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.095695 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-s6zwp"] Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.126216 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-rw942"] Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.140424 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:46 crc kubenswrapper[4813]: E0317 09:13:46.141132 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:46.641083942 +0000 UTC m=+248.741887451 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.141364 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" event={"ID":"364cd6ac-033f-455e-a95a-10d50f055fb2","Type":"ContainerStarted","Data":"c78ddaea266e5875143f775580e916596cacb802d70464b450d2ed17fea803df"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.172277 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" event={"ID":"74f92156-f803-433a-99a4-981202076e81","Type":"ContainerStarted","Data":"fca401c8c10481f6181cc4f34ed269a2838d6d313adba4fdfe6a786a51f43e33"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.245912 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.245948 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wcd9z" event={"ID":"72fa1c68-872b-45a3-98b7-b0b1ebc08ad7","Type":"ContainerStarted","Data":"e2acdcd1aadcff06727e99851be5531e4b8c2262c20a497c3ed5e07518f6e00a"} Mar 17 09:13:46 crc kubenswrapper[4813]: E0317 09:13:46.246443 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:46.74643113 +0000 UTC m=+248.847234629 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.253475 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-kns96" podStartSLOduration=184.25345773 podStartE2EDuration="3m4.25345773s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:46.252268319 +0000 UTC m=+248.353071818" watchObservedRunningTime="2026-03-17 09:13:46.25345773 +0000 UTC m=+248.354261229" Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.293863 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" podStartSLOduration=184.293849104 podStartE2EDuration="3m4.293849104s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:46.291328859 +0000 UTC m=+248.392132358" watchObservedRunningTime="2026-03-17 09:13:46.293849104 +0000 UTC m=+248.394652603" Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.295546 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-ghwhk"] Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.305858 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx" event={"ID":"9902dff6-4484-44ba-af31-877293debb91","Type":"ContainerStarted","Data":"279f6b09a70acffa63153e93e5fccd847587f484ea81f24fecf61b8bbad711bb"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.316469 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562312-hjpcp"] Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.323897 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-cjn8k"] Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.326019 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-m97th" podStartSLOduration=184.32600137 podStartE2EDuration="3m4.32600137s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:46.32336514 +0000 UTC m=+248.424168639" watchObservedRunningTime="2026-03-17 09:13:46.32600137 +0000 UTC m=+248.426804869" Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.346676 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:46 crc kubenswrapper[4813]: E0317 09:13:46.347632 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:46.847612646 +0000 UTC m=+248.948416145 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.348874 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8qg9n"] Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.361914 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4rqzj" event={"ID":"b828a0e2-3def-4bb5-8f54-2792376423cd","Type":"ContainerStarted","Data":"990d97359c2ef31bb1acb4f272f1d98c246a3682f4b5202adc9ef44238f60a84"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.361953 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4rqzj" Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.363739 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bpfb" event={"ID":"03c2477b-c56a-4354-bf4f-170b83f9240f","Type":"ContainerStarted","Data":"90846f7b9280325cc68c9630d850c81c10a4aaabe97511e5467ed09ba499d85c"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.384556 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-njr98" event={"ID":"93cb7b2d-1c40-40be-82d7-1858272444ec","Type":"ContainerStarted","Data":"ed70d5650c389e08e3b9e1ba704089aff5f588875472cef0f8412ed664e4a90c"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.431769 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7tdh4" event={"ID":"58e4f2e1-b751-41e1-a56e-b62fd82b7caf","Type":"ContainerStarted","Data":"58d96a4f3b77cc079a04e120af2e3fe6b99ba21096f3f334981d5ed845852e30"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.448316 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:46 crc kubenswrapper[4813]: E0317 09:13:46.449637 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:46.94962131 +0000 UTC m=+249.050424929 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.451904 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.463070 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lrhrs" event={"ID":"f253587f-9f27-474d-8f46-21f1e6a79d97","Type":"ContainerStarted","Data":"bb5dcbe1e5984030d34b660f50fe5309a963872483cfb9311349bdc0a1659ab3"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.477428 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s2hhg" event={"ID":"d32fb60f-2d76-4e02-9839-535e1543eff7","Type":"ContainerStarted","Data":"021c6019434de7cdb9932ffe177d231836d63250d7e14d1bc79348c106916329"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.502330 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5djq" event={"ID":"37f3773a-2367-411b-b8e4-c4ae96970de3","Type":"ContainerStarted","Data":"6e23faa7665638136b6f9c73445e1b61cb0975b28392ffef16f5226c391f6536"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.531068 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gczkp" event={"ID":"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b","Type":"ContainerStarted","Data":"94128f0b645b155430c88702e32c32b67ef180b587a3dcca8423aad4abcc936d"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.550086 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:46 crc kubenswrapper[4813]: E0317 09:13:46.551221 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:47.051203109 +0000 UTC m=+249.152006608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.567484 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dpghh" event={"ID":"845ee13b-7ed9-4910-958b-5269812237f0","Type":"ContainerStarted","Data":"d9c36c9b788f6fcb1026559c1bd08f83ca0f55ffb1f3b473955d0fda7742ebbf"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.567533 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dpghh" event={"ID":"845ee13b-7ed9-4910-958b-5269812237f0","Type":"ContainerStarted","Data":"9659c32c94d48c65f3ea66ab02d14dde7aa24ab6efa6213423119b1912004113"} Mar 17 09:13:46 crc kubenswrapper[4813]: W0317 09:13:46.572899 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf760de2f_292d_48fe_ac0c_a97aa37f556a.slice/crio-7c4ef7c6020c39f4e92dc63aaf43ea34de8202f77669f7d7f7880f36ee71f100 WatchSource:0}: Error finding container 7c4ef7c6020c39f4e92dc63aaf43ea34de8202f77669f7d7f7880f36ee71f100: Status 404 returned error can't find the container with id 7c4ef7c6020c39f4e92dc63aaf43ea34de8202f77669f7d7f7880f36ee71f100 Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.577153 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-tzt6n" event={"ID":"550c18c5-f4e0-4da1-9692-c6cacfa882af","Type":"ContainerStarted","Data":"2d4e9893e7812f03c0e8890fc4c65fdbc3a2d3655348120ce682560c3705d8ac"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.586303 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" podStartSLOduration=183.586287904 podStartE2EDuration="3m3.586287904s" podCreationTimestamp="2026-03-17 09:10:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:46.537063547 +0000 UTC m=+248.637867056" watchObservedRunningTime="2026-03-17 09:13:46.586287904 +0000 UTC m=+248.687091403" Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.586939 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-6lsz5"] Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.609413 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8hxbx" event={"ID":"ed67dcc9-2a14-4e28-86df-da77a8d20012","Type":"ContainerStarted","Data":"36b2c7dd60dc26b44c6ba93d96b2d51e364c7ed67b20323331c5876b8af6b694"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.612589 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-wttgv" event={"ID":"000e63dc-8228-486c-a6ee-12843810b5af","Type":"ContainerStarted","Data":"0e408a419bc564cf8e21aed7b7199aa10d8ebd967dc005c60c782c413ad4984f"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.616795 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-xb4k9" event={"ID":"3aa0fc32-ed11-482b-9a6d-fcd59544bf8d","Type":"ContainerStarted","Data":"5f2a3af66e994d421f9e46466a63fdd84eb30b09966a56680ee4d7ca97d76eea"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.617659 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-xb4k9" Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.618782 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwbwr" event={"ID":"2bd4438d-8a23-4e8b-aad7-f5775f1d88a8","Type":"ContainerStarted","Data":"3fb035466fae238a0197a2857d644ad21dde0d3b7404a5b12c27574625893559"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.623857 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-xb4k9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.623896 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xb4k9" podUID="3aa0fc32-ed11-482b-9a6d-fcd59544bf8d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.624720 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2w9vb" event={"ID":"f7603094-bb1d-4d77-b463-2270003b2805","Type":"ContainerStarted","Data":"2967cb5bd44fd0526a4158a5226c69df2a26093d950f605e61c3ba21d8a5f068"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.640577 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-qwlbl" event={"ID":"94f282fa-7b22-41e3-9911-073b05d41200","Type":"ContainerStarted","Data":"9e83d12e0cb648d54fc684b8eebbb051c735c0472d6a0708b530175d04e0b3d5"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.641245 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" podStartSLOduration=184.641234395 podStartE2EDuration="3m4.641234395s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:46.60145189 +0000 UTC m=+248.702255389" watchObservedRunningTime="2026-03-17 09:13:46.641234395 +0000 UTC m=+248.742037894" Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.655026 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2k58b" podStartSLOduration=184.655006694 podStartE2EDuration="3m4.655006694s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:46.641049348 +0000 UTC m=+248.741852847" watchObservedRunningTime="2026-03-17 09:13:46.655006694 +0000 UTC m=+248.755810193" Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.656412 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:46 crc kubenswrapper[4813]: E0317 09:13:46.657500 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:47.157486828 +0000 UTC m=+249.258290327 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.662026 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7w9l5" event={"ID":"3abdbcf2-c3d0-4049-b08b-45fa02672dd4","Type":"ContainerStarted","Data":"724e805903665723ab61c02f51e658019e709edd616f76123a7c8462a9ba90fe"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.668339 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kpxlg" event={"ID":"12efeac2-9fc3-425a-8874-3c2d957e04ca","Type":"ContainerStarted","Data":"c3b97075d5de6972ce7221dba0f3a7d6e05716da9479ff5a04ba771aaddebe2e"} Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.697713 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.700237 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-2fmh2" Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.701647 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.729516 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.758374 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-2fmh2" podStartSLOduration=184.758353603 podStartE2EDuration="3m4.758353603s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:46.732821544 +0000 UTC m=+248.833625043" watchObservedRunningTime="2026-03-17 09:13:46.758353603 +0000 UTC m=+248.859157122" Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.760193 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:46 crc kubenswrapper[4813]: E0317 09:13:46.761495 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:47.261473439 +0000 UTC m=+249.362276938 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.788858 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5djq" podStartSLOduration=184.788842002 podStartE2EDuration="3m4.788842002s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:46.759812723 +0000 UTC m=+248.860616222" watchObservedRunningTime="2026-03-17 09:13:46.788842002 +0000 UTC m=+248.889645501" Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.789132 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-wcd9z" podStartSLOduration=184.789127971 podStartE2EDuration="3m4.789127971s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:46.78820758 +0000 UTC m=+248.889011079" watchObservedRunningTime="2026-03-17 09:13:46.789127971 +0000 UTC m=+248.889931470" Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.862334 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:46 crc kubenswrapper[4813]: E0317 09:13:46.862876 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:47.362865782 +0000 UTC m=+249.463669282 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.879121 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lrhrs" podStartSLOduration=184.879103775 podStartE2EDuration="3m4.879103775s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:46.877421658 +0000 UTC m=+248.978225157" watchObservedRunningTime="2026-03-17 09:13:46.879103775 +0000 UTC m=+248.979907274" Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.918342 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-tzt6n" podStartSLOduration=184.918324371 podStartE2EDuration="3m4.918324371s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:46.91651467 +0000 UTC m=+249.017318169" watchObservedRunningTime="2026-03-17 09:13:46.918324371 +0000 UTC m=+249.019127870" Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.958482 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4rqzj" podStartSLOduration=184.958467528 podStartE2EDuration="3m4.958467528s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:46.956389198 +0000 UTC m=+249.057192697" watchObservedRunningTime="2026-03-17 09:13:46.958467528 +0000 UTC m=+249.059271027" Mar 17 09:13:46 crc kubenswrapper[4813]: I0317 09:13:46.965552 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:46 crc kubenswrapper[4813]: E0317 09:13:46.965881 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:47.46586658 +0000 UTC m=+249.566670079 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.016482 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-s2hhg" podStartSLOduration=185.016467623 podStartE2EDuration="3m5.016467623s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:46.986559245 +0000 UTC m=+249.087362744" watchObservedRunningTime="2026-03-17 09:13:47.016467623 +0000 UTC m=+249.117271122" Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.059134 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-wwbwr" podStartSLOduration=185.059118446 podStartE2EDuration="3m5.059118446s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:47.017714986 +0000 UTC m=+249.118518485" watchObservedRunningTime="2026-03-17 09:13:47.059118446 +0000 UTC m=+249.159921945" Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.069484 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:47 crc kubenswrapper[4813]: E0317 09:13:47.069837 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:47.5698258 +0000 UTC m=+249.670629299 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.100444 4813 patch_prober.go:28] interesting pod/router-default-5444994796-m97th container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 09:13:47 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Mar 17 09:13:47 crc kubenswrapper[4813]: [+]process-running ok Mar 17 09:13:47 crc kubenswrapper[4813]: healthz check failed Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.100484 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-m97th" podUID="68399978-6f43-4458-a1e1-58d5984de8f5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.125431 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-njr98" podStartSLOduration=185.125417843 podStartE2EDuration="3m5.125417843s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:47.124650698 +0000 UTC m=+249.225454197" watchObservedRunningTime="2026-03-17 09:13:47.125417843 +0000 UTC m=+249.226221342" Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.168325 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-2w9vb" podStartSLOduration=185.168309114 podStartE2EDuration="3m5.168309114s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:47.166297086 +0000 UTC m=+249.267100595" watchObservedRunningTime="2026-03-17 09:13:47.168309114 +0000 UTC m=+249.269112613" Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.177366 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:47 crc kubenswrapper[4813]: E0317 09:13:47.177826 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:47.677809198 +0000 UTC m=+249.778612697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.196545 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kpxlg" podStartSLOduration=185.196530026 podStartE2EDuration="3m5.196530026s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:47.194631601 +0000 UTC m=+249.295435100" watchObservedRunningTime="2026-03-17 09:13:47.196530026 +0000 UTC m=+249.297333525" Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.232206 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-xb4k9" podStartSLOduration=185.23219003 podStartE2EDuration="3m5.23219003s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:47.229629453 +0000 UTC m=+249.330432972" watchObservedRunningTime="2026-03-17 09:13:47.23219003 +0000 UTC m=+249.332993529" Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.258234 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-8hxbx" podStartSLOduration=6.258214386 podStartE2EDuration="6.258214386s" podCreationTimestamp="2026-03-17 09:13:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:47.255211634 +0000 UTC m=+249.356015133" watchObservedRunningTime="2026-03-17 09:13:47.258214386 +0000 UTC m=+249.359017885" Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.278571 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:47 crc kubenswrapper[4813]: E0317 09:13:47.278919 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:47.77890676 +0000 UTC m=+249.879710259 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.390456 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:47 crc kubenswrapper[4813]: E0317 09:13:47.390951 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:47.890935275 +0000 UTC m=+249.991738784 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.505521 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:47 crc kubenswrapper[4813]: E0317 09:13:47.506171 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:48.006156389 +0000 UTC m=+250.106959888 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.605205 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.605241 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.607852 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:47 crc kubenswrapper[4813]: E0317 09:13:47.608632 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:48.108609819 +0000 UTC m=+250.209413308 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.679220 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7tdh4" event={"ID":"58e4f2e1-b751-41e1-a56e-b62fd82b7caf","Type":"ContainerStarted","Data":"88b8d959fb3f5130f51d9c24e00e135f6a098453a33b0f26bee417388952d39a"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.711778 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tbl5t" event={"ID":"7954ca65-b7a4-47de-bc7c-2f57f0fee19a","Type":"ContainerStarted","Data":"7b9b708a109ba082775f5072f27081ada269d423e96527770a3f0f47f2bb6b53"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.711836 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tbl5t" event={"ID":"7954ca65-b7a4-47de-bc7c-2f57f0fee19a","Type":"ContainerStarted","Data":"f952134ef0079ce56a94bcb8fc4027f439998431fff062a1e92e5c959ac0b643"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.712278 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:47 crc kubenswrapper[4813]: E0317 09:13:47.712730 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:48.212716315 +0000 UTC m=+250.313519824 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.724185 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8qg9n" event={"ID":"f760de2f-292d-48fe-ac0c-a97aa37f556a","Type":"ContainerStarted","Data":"7c4ef7c6020c39f4e92dc63aaf43ea34de8202f77669f7d7f7880f36ee71f100"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.735092 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6lsz5" event={"ID":"173f4a11-926b-4cc0-a70c-c369ac37bac2","Type":"ContainerStarted","Data":"e036620d70e57c355c2dfcb6a5093810b9cab2fc65400c43f7af455480efc3f3"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.735951 4813 patch_prober.go:28] interesting pod/apiserver-76f77b778f-kns96 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 17 09:13:47 crc kubenswrapper[4813]: [+]log ok Mar 17 09:13:47 crc kubenswrapper[4813]: [+]etcd ok Mar 17 09:13:47 crc kubenswrapper[4813]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 17 09:13:47 crc kubenswrapper[4813]: [+]poststarthook/generic-apiserver-start-informers ok Mar 17 09:13:47 crc kubenswrapper[4813]: [+]poststarthook/max-in-flight-filter ok Mar 17 09:13:47 crc kubenswrapper[4813]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 17 09:13:47 crc kubenswrapper[4813]: [+]poststarthook/image.openshift.io-apiserver-caches ok Mar 17 09:13:47 crc kubenswrapper[4813]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Mar 17 09:13:47 crc kubenswrapper[4813]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Mar 17 09:13:47 crc kubenswrapper[4813]: [+]poststarthook/project.openshift.io-projectcache ok Mar 17 09:13:47 crc kubenswrapper[4813]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Mar 17 09:13:47 crc kubenswrapper[4813]: [+]poststarthook/openshift.io-startinformers ok Mar 17 09:13:47 crc kubenswrapper[4813]: [+]poststarthook/openshift.io-restmapperupdater ok Mar 17 09:13:47 crc kubenswrapper[4813]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Mar 17 09:13:47 crc kubenswrapper[4813]: livez check failed Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.735994 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-kns96" podUID="26a86c9b-1645-49c2-a2fb-78b210984363" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.762959 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" event={"ID":"eb147d37-445a-48f2-a2e6-fb16f3575529","Type":"ContainerStarted","Data":"784fe43f2fe001d814ee0d2a39bf5a84db11b124157af1867d559d10ec566075"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.783245 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" podStartSLOduration=185.783230885 podStartE2EDuration="3m5.783230885s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:47.783043659 +0000 UTC m=+249.883847158" watchObservedRunningTime="2026-03-17 09:13:47.783230885 +0000 UTC m=+249.884034384" Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.799917 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7w9l5" event={"ID":"3abdbcf2-c3d0-4049-b08b-45fa02672dd4","Type":"ContainerStarted","Data":"dd9d015dc7452873cecd074d17a9c6165d95f0429b6423036487fca99ea2d8cd"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.799960 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7w9l5" event={"ID":"3abdbcf2-c3d0-4049-b08b-45fa02672dd4","Type":"ContainerStarted","Data":"39b6cb25f023462bb65402746c26ad450f1d0b0e15840c0a07ce09e127a8a654"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.800807 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7w9l5" Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.813396 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:47 crc kubenswrapper[4813]: E0317 09:13:47.813808 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:48.313792166 +0000 UTC m=+250.414595665 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.822953 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-kpxlg" event={"ID":"12efeac2-9fc3-425a-8874-3c2d957e04ca","Type":"ContainerStarted","Data":"a752bafaa0ab6072a7284ecc9435723c243909469552f2e85aa321922ab24574"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.831724 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7w9l5" podStartSLOduration=185.831702196 podStartE2EDuration="3m5.831702196s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:47.830220576 +0000 UTC m=+249.931024095" watchObservedRunningTime="2026-03-17 09:13:47.831702196 +0000 UTC m=+249.932505695" Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.855166 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-qwlbl" event={"ID":"94f282fa-7b22-41e3-9911-073b05d41200","Type":"ContainerStarted","Data":"5873e054897787e7506587d55c9222315b73fc0703274ee8e8e830c654ddf3af"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.858246 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-wttgv" event={"ID":"000e63dc-8228-486c-a6ee-12843810b5af","Type":"ContainerStarted","Data":"e52337759d31f0a568bc62cb97d51866097afb8e2046670287122d80bca73548"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.860855 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vm474" event={"ID":"1507ff6e-1511-4c90-998e-1fb84da97e5b","Type":"ContainerStarted","Data":"f8f3b5183145b470969e4f56967cda0427481713610d3de90f2414872a48da83"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.882933 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-wttgv" podStartSLOduration=184.882912991 podStartE2EDuration="3m4.882912991s" podCreationTimestamp="2026-03-17 09:10:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:47.88174516 +0000 UTC m=+249.982548729" watchObservedRunningTime="2026-03-17 09:13:47.882912991 +0000 UTC m=+249.983716500" Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.891407 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-j5djq" event={"ID":"37f3773a-2367-411b-b8e4-c4ae96970de3","Type":"ContainerStarted","Data":"5e8679cdd8817940a20852f555e9e6655ea6fac5aa8b0af344f28cf00095926d"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.910510 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx" event={"ID":"9902dff6-4484-44ba-af31-877293debb91","Type":"ContainerStarted","Data":"a4bfc9c6b5a694c3953c1d6864c3136e14fb013a9eac9a671b074fb5cd984f18"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.911167 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx" Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.913390 4813 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-slqwx container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.42:5443/healthz\": dial tcp 10.217.0.42:5443: connect: connection refused" start-of-body= Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.913429 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx" podUID="9902dff6-4484-44ba-af31-877293debb91" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.42:5443/healthz\": dial tcp 10.217.0.42:5443: connect: connection refused" Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.914386 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:47 crc kubenswrapper[4813]: E0317 09:13:47.915831 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:48.415812871 +0000 UTC m=+250.516616430 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.917917 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7" event={"ID":"b3992349-bc61-405e-be75-609cce2fad10","Type":"ContainerStarted","Data":"6abc6fdbecc5d29e40bf98aa15092c6a91d5d8667c098322c7987cf7b1896af4"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.917953 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7" event={"ID":"b3992349-bc61-405e-be75-609cce2fad10","Type":"ContainerStarted","Data":"44248ad0ada19b5a9f1ab05fa4dc5f199ea2c9e424f5d54e750f54345d728673"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.919884 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tjvgc" event={"ID":"0021cb4b-86a2-46a5-83f6-109fee3e517b","Type":"ContainerStarted","Data":"aa0d9781ddcad420ffdda5635a86b0f4840dc2062da849ee6b1e0f547d33088d"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.919965 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tjvgc" event={"ID":"0021cb4b-86a2-46a5-83f6-109fee3e517b","Type":"ContainerStarted","Data":"864576cee18e39e19d535d5e4ca49f255fb3d34c427573a93ef7674bf564a723"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.920235 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tjvgc" Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.922054 4813 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-tjvgc container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" start-of-body= Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.922095 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tjvgc" podUID="0021cb4b-86a2-46a5-83f6-109fee3e517b" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.43:8443/healthz\": dial tcp 10.217.0.43:8443: connect: connection refused" Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.922322 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-xb4k9" event={"ID":"3aa0fc32-ed11-482b-9a6d-fcd59544bf8d","Type":"ContainerStarted","Data":"b506d6b8a3793b3ec00bb1384e7e4729a0da3642f766c2c7e499cdfe506e72a2"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.923490 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-xb4k9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.923514 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xb4k9" podUID="3aa0fc32-ed11-482b-9a6d-fcd59544bf8d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.924504 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-8hxbx" event={"ID":"ed67dcc9-2a14-4e28-86df-da77a8d20012","Type":"ContainerStarted","Data":"e74c44be19681fca89fad6a03368fbd234c8c541dc659bf57442434e6783b9ea"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.927413 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562312-hjpcp" event={"ID":"8626af43-6b5e-49e1-b19f-3986cdecdd2d","Type":"ContainerStarted","Data":"13518e37a85cf60fdf0ccedf00537690826be4b4ec70bbb11418dd03a067aa1c"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.932511 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ghwhk" event={"ID":"d89e7afc-8f04-4996-b087-306543ef5acb","Type":"ContainerStarted","Data":"804a1c20736a6c97ead4e6e9bcd9df026c169c225924d99b1562ad9b4bd0ca9e"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.932553 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ghwhk" event={"ID":"d89e7afc-8f04-4996-b087-306543ef5acb","Type":"ContainerStarted","Data":"dca9e45084669de8503081006fd75c72a182f798276cc625018095c289666711"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.948872 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-cjn8k" event={"ID":"da36625d-0c68-48b8-9608-c7dedd304064","Type":"ContainerStarted","Data":"cfc552a5f31b889c4c4a3e56934b8f8ec844b98b11dd0367db48916d4f77643c"} Mar 17 09:13:47 crc kubenswrapper[4813]: I0317 09:13:47.991740 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dpghh" event={"ID":"845ee13b-7ed9-4910-958b-5269812237f0","Type":"ContainerStarted","Data":"cf1310c73201d66835ff895a2fb6c8982a761b56dfb62c6adb68dfced4eca0bc"} Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.000996 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vm474" podStartSLOduration=186.000981612 podStartE2EDuration="3m6.000981612s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:47.949099134 +0000 UTC m=+250.049902633" watchObservedRunningTime="2026-03-17 09:13:48.000981612 +0000 UTC m=+250.101785101" Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.018201 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:48 crc kubenswrapper[4813]: E0317 09:13:48.018997 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:48.518979034 +0000 UTC m=+250.619782533 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.023057 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" event={"ID":"74f92156-f803-433a-99a4-981202076e81","Type":"ContainerStarted","Data":"f6ab163fea4b95b91341d3e6ebaee0ae1e7c34b859dfe7cfc8c1d4b4c0e83595"} Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.024008 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.041655 4813 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5wl9g container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.041710 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" podUID="74f92156-f803-433a-99a4-981202076e81" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.043008 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-rw942" event={"ID":"a6f67978-6c47-4927-ae94-1fef4eee7b18","Type":"ContainerStarted","Data":"9f47bd30dc2ec112010d80ea00bdf5a1c2e4e139201f534f975698b95f4e220e"} Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.043059 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-rw942" event={"ID":"a6f67978-6c47-4927-ae94-1fef4eee7b18","Type":"ContainerStarted","Data":"e060a42c9d5d2841926986f33a7848e6f10344472d74adae827929c533cd9dc9"} Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.053118 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx" podStartSLOduration=186.053103396 podStartE2EDuration="3m6.053103396s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:48.00154558 +0000 UTC m=+250.102349089" watchObservedRunningTime="2026-03-17 09:13:48.053103396 +0000 UTC m=+250.153906895" Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.068386 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-lrhrs" event={"ID":"f253587f-9f27-474d-8f46-21f1e6a79d97","Type":"ContainerStarted","Data":"2db62567a7ffd4c1a5d96d1dda8fca46fda55889c02a59a87bdd821ff005ba0b"} Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.100028 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bpfb" event={"ID":"03c2477b-c56a-4354-bf4f-170b83f9240f","Type":"ContainerStarted","Data":"f192137e8fd5d8653486e95bb7e3e53157361959102cb9e632bc28414be348a5"} Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.103796 4813 patch_prober.go:28] interesting pod/router-default-5444994796-m97th container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 09:13:48 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Mar 17 09:13:48 crc kubenswrapper[4813]: [+]process-running ok Mar 17 09:13:48 crc kubenswrapper[4813]: healthz check failed Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.103851 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-m97th" podUID="68399978-6f43-4458-a1e1-58d5984de8f5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.116939 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tjvgc" podStartSLOduration=186.116919969 podStartE2EDuration="3m6.116919969s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:48.063694627 +0000 UTC m=+250.164498126" watchObservedRunningTime="2026-03-17 09:13:48.116919969 +0000 UTC m=+250.217723478" Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.118610 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2w9vb" event={"ID":"f7603094-bb1d-4d77-b463-2270003b2805","Type":"ContainerStarted","Data":"011538bd3ac194eac4eaac10f92db465d10f8ff49e5ec1b4f73b7400a510e831"} Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.124392 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:48 crc kubenswrapper[4813]: E0317 09:13:48.126515 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:48.626504676 +0000 UTC m=+250.727308175 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.161568 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7" podStartSLOduration=186.161552379 podStartE2EDuration="3m6.161552379s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:48.160970999 +0000 UTC m=+250.261774498" watchObservedRunningTime="2026-03-17 09:13:48.161552379 +0000 UTC m=+250.262355878" Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.161971 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-ghwhk" podStartSLOduration=185.161965873 podStartE2EDuration="3m5.161965873s" podCreationTimestamp="2026-03-17 09:10:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:48.118157622 +0000 UTC m=+250.218961121" watchObservedRunningTime="2026-03-17 09:13:48.161965873 +0000 UTC m=+250.262769372" Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.174582 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" event={"ID":"364cd6ac-033f-455e-a95a-10d50f055fb2","Type":"ContainerStarted","Data":"18ac861d6e271d6db16e47b37ddfef57c778b764d125d8a8891f72eb4c9a5482"} Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.238320 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-2bpfb" podStartSLOduration=186.238305703 podStartE2EDuration="3m6.238305703s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:48.237018439 +0000 UTC m=+250.337821938" watchObservedRunningTime="2026-03-17 09:13:48.238305703 +0000 UTC m=+250.339109202" Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.238948 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-rw942" podStartSLOduration=7.238942755 podStartE2EDuration="7.238942755s" podCreationTimestamp="2026-03-17 09:13:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:48.193647623 +0000 UTC m=+250.294451132" watchObservedRunningTime="2026-03-17 09:13:48.238942755 +0000 UTC m=+250.339746254" Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.240096 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:48 crc kubenswrapper[4813]: E0317 09:13:48.241023 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:48.741007035 +0000 UTC m=+250.841810534 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.273447 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s6zwp" event={"ID":"f7c4987b-f8f6-4d39-b7f1-6f80c1384293","Type":"ContainerStarted","Data":"1f2586fc8db7839b67b7598ada6153ee01ead032174d875f32ca7d69f9734493"} Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.290938 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dpghh" podStartSLOduration=186.290922135 podStartE2EDuration="3m6.290922135s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:48.290637435 +0000 UTC m=+250.391440934" watchObservedRunningTime="2026-03-17 09:13:48.290922135 +0000 UTC m=+250.391725634" Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.323148 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" podStartSLOduration=186.323126881 podStartE2EDuration="3m6.323126881s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:48.322176519 +0000 UTC m=+250.422980048" watchObservedRunningTime="2026-03-17 09:13:48.323126881 +0000 UTC m=+250.423930380" Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.352926 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:48 crc kubenswrapper[4813]: E0317 09:13:48.354190 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:48.854177269 +0000 UTC m=+250.954980768 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.454477 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:48 crc kubenswrapper[4813]: E0317 09:13:48.456276 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:48.956254785 +0000 UTC m=+251.057058284 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.556511 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:48 crc kubenswrapper[4813]: E0317 09:13:48.557033 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:49.057020317 +0000 UTC m=+251.157823816 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.579466 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.579510 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.587725 4813 patch_prober.go:28] interesting pod/apiserver-7bbb656c7d-h87sm container/oauth-apiserver namespace/openshift-oauth-apiserver: Startup probe status=failure output="Get \"https://10.217.0.20:8443/livez\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.587792 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" podUID="eb147d37-445a-48f2-a2e6-fb16f3575529" containerName="oauth-apiserver" probeResult="failure" output="Get \"https://10.217.0.20:8443/livez\": dial tcp 10.217.0.20:8443: connect: connection refused" Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.649984 4813 ???:1] "http: TLS handshake error from 192.168.126.11:52188: no serving certificate available for the kubelet" Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.657949 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:48 crc kubenswrapper[4813]: E0317 09:13:48.658683 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:49.158664019 +0000 UTC m=+251.259467528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.751300 4813 ???:1] "http: TLS handshake error from 192.168.126.11:52190: no serving certificate available for the kubelet" Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.760385 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:48 crc kubenswrapper[4813]: E0317 09:13:48.760878 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:49.260864169 +0000 UTC m=+251.361667668 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.766842 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-jzbwp" podStartSLOduration=186.766814642 podStartE2EDuration="3m6.766814642s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:48.351761707 +0000 UTC m=+250.452565206" watchObservedRunningTime="2026-03-17 09:13:48.766814642 +0000 UTC m=+250.867618141" Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.835166 4813 ???:1] "http: TLS handshake error from 192.168.126.11:52202: no serving certificate available for the kubelet" Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.861811 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:48 crc kubenswrapper[4813]: E0317 09:13:48.862012 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:49.361982732 +0000 UTC m=+251.462786241 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.862266 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:48 crc kubenswrapper[4813]: E0317 09:13:48.862555 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:49.362543642 +0000 UTC m=+251.463347141 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.964771 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:48 crc kubenswrapper[4813]: E0317 09:13:48.965093 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:49.465078853 +0000 UTC m=+251.565882352 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:48 crc kubenswrapper[4813]: I0317 09:13:48.979953 4813 ???:1] "http: TLS handshake error from 192.168.126.11:52212: no serving certificate available for the kubelet" Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.066351 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:49 crc kubenswrapper[4813]: E0317 09:13:49.066762 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:49.566746516 +0000 UTC m=+251.667550015 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.089931 4813 patch_prober.go:28] interesting pod/router-default-5444994796-m97th container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 09:13:49 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Mar 17 09:13:49 crc kubenswrapper[4813]: [+]process-running ok Mar 17 09:13:49 crc kubenswrapper[4813]: healthz check failed Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.090023 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-m97th" podUID="68399978-6f43-4458-a1e1-58d5984de8f5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.094837 4813 ???:1] "http: TLS handshake error from 192.168.126.11:52214: no serving certificate available for the kubelet" Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.168024 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:49 crc kubenswrapper[4813]: E0317 09:13:49.168146 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:49.668129239 +0000 UTC m=+251.768932738 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.168284 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:49 crc kubenswrapper[4813]: E0317 09:13:49.168587 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:49.668579703 +0000 UTC m=+251.769383202 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.202463 4813 ???:1] "http: TLS handshake error from 192.168.126.11:52220: no serving certificate available for the kubelet" Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.269051 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:49 crc kubenswrapper[4813]: E0317 09:13:49.269191 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:49.76917325 +0000 UTC m=+251.869976739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.269335 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:49 crc kubenswrapper[4813]: E0317 09:13:49.269673 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:49.769665136 +0000 UTC m=+251.870468635 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.277472 4813 patch_prober.go:28] interesting pod/openshift-config-operator-7777fb866f-4rqzj container/openshift-config-operator namespace/openshift-config-operator: Readiness probe status=failure output="Get \"https://10.217.0.12:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.277530 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4rqzj" podUID="b828a0e2-3def-4bb5-8f54-2792376423cd" containerName="openshift-config-operator" probeResult="failure" output="Get \"https://10.217.0.12:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.290015 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-vm474" event={"ID":"1507ff6e-1511-4c90-998e-1fb84da97e5b","Type":"ContainerStarted","Data":"7e43ac64c2017414afef2d5e87478f5fba39b8c34bff2def7437be43ccf6408e"} Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.295017 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gczkp" event={"ID":"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b","Type":"ContainerStarted","Data":"f30f9d2aeaa790ea4364ec4cb256a14d50841961e981248931ab7a7360d2cc57"} Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.302948 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-qwlbl" event={"ID":"94f282fa-7b22-41e3-9911-073b05d41200","Type":"ContainerStarted","Data":"abe23d682c12ef16b481b0a9d2959d6eefd6ac295d80004bca56037a295bc4c4"} Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.307709 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-cjn8k" event={"ID":"da36625d-0c68-48b8-9608-c7dedd304064","Type":"ContainerStarted","Data":"35f55cf908e89f77a330c0ac5cbf03e205d0957f536c5d1d70a883130d7bfeec"} Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.307753 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-cjn8k" event={"ID":"da36625d-0c68-48b8-9608-c7dedd304064","Type":"ContainerStarted","Data":"ed0935267cd90bf175f1988900b2b97a49628c080f761643920c9274d7dc3a5e"} Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.308252 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-cjn8k" Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.309859 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tbl5t" event={"ID":"7954ca65-b7a4-47de-bc7c-2f57f0fee19a","Type":"ContainerStarted","Data":"9dd01913ef6186155be9a0c06b739228b1e14e5ee073fdd8b25d14cdcf1ab758"} Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.325464 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7tdh4" event={"ID":"58e4f2e1-b751-41e1-a56e-b62fd82b7caf","Type":"ContainerStarted","Data":"d857421fbdea9346026f0cb2721e15d5e4fb90d41dc9a26bda378780bca236eb"} Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.328877 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6lsz5" event={"ID":"173f4a11-926b-4cc0-a70c-c369ac37bac2","Type":"ContainerStarted","Data":"5bc5518c569ea38a18ade8856a4afdc5027abba67345a997efc1c51293f7edf5"} Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.328955 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-6lsz5" event={"ID":"173f4a11-926b-4cc0-a70c-c369ac37bac2","Type":"ContainerStarted","Data":"72d284005edf0e6fa044065e49e6f6b258e538ae9da13db43163ea21387396af"} Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.338884 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s6zwp" event={"ID":"f7c4987b-f8f6-4d39-b7f1-6f80c1384293","Type":"ContainerStarted","Data":"fd83fe4d49e98cf02e72c2fb5dbe94f42d4610743b9257bfd9bbca126fdf73c6"} Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.338938 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s6zwp" event={"ID":"f7c4987b-f8f6-4d39-b7f1-6f80c1384293","Type":"ContainerStarted","Data":"82f820894e9a088733a56586cd72a7164f13c248c5ca6ba4a81471cfa3554425"} Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.339062 4813 ???:1] "http: TLS handshake error from 192.168.126.11:52228: no serving certificate available for the kubelet" Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.351042 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8qg9n" event={"ID":"f760de2f-292d-48fe-ac0c-a97aa37f556a","Type":"ContainerStarted","Data":"5ae4bc7b41653c78c750ca2021e8281e8d8da62a085ff135020e913ea26a8f62"} Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.351118 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8qg9n" Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.357162 4813 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-8qg9n container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" start-of-body= Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.357219 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8qg9n" podUID="f760de2f-292d-48fe-ac0c-a97aa37f556a" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.24:8443/healthz\": dial tcp 10.217.0.24:8443: connect: connection refused" Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.360677 4813 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-5wl9g container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.360717 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" podUID="74f92156-f803-433a-99a4-981202076e81" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.360771 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-xb4k9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.360788 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xb4k9" podUID="3aa0fc32-ed11-482b-9a6d-fcd59544bf8d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.373413 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:49 crc kubenswrapper[4813]: E0317 09:13:49.373784 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:49.87374605 +0000 UTC m=+251.974549549 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.374178 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:49 crc kubenswrapper[4813]: E0317 09:13:49.374638 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:49.87461883 +0000 UTC m=+251.975422329 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.383702 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-tjvgc" Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.386523 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-tbl5t" podStartSLOduration=187.386506125 podStartE2EDuration="3m7.386506125s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:49.384042672 +0000 UTC m=+251.484846171" watchObservedRunningTime="2026-03-17 09:13:49.386506125 +0000 UTC m=+251.487309624" Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.407406 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-qwlbl" podStartSLOduration=187.407383456 podStartE2EDuration="3m7.407383456s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:49.342258328 +0000 UTC m=+251.443061827" watchObservedRunningTime="2026-03-17 09:13:49.407383456 +0000 UTC m=+251.508186955" Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.434872 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-7tdh4" podStartSLOduration=187.434857641 podStartE2EDuration="3m7.434857641s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:49.432634446 +0000 UTC m=+251.533437945" watchObservedRunningTime="2026-03-17 09:13:49.434857641 +0000 UTC m=+251.535661140" Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.445499 4813 ???:1] "http: TLS handshake error from 192.168.126.11:52240: no serving certificate available for the kubelet" Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.466487 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-6lsz5" podStartSLOduration=187.466469098 podStartE2EDuration="3m7.466469098s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:49.465247217 +0000 UTC m=+251.566050716" watchObservedRunningTime="2026-03-17 09:13:49.466469098 +0000 UTC m=+251.567272597" Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.479638 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:49 crc kubenswrapper[4813]: E0317 09:13:49.481383 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:49.981368366 +0000 UTC m=+252.082171865 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.504873 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-cjn8k" podStartSLOduration=8.504856995 podStartE2EDuration="8.504856995s" podCreationTimestamp="2026-03-17 09:13:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:49.502528186 +0000 UTC m=+251.603331685" watchObservedRunningTime="2026-03-17 09:13:49.504856995 +0000 UTC m=+251.605660494" Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.558554 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-s6zwp" podStartSLOduration=187.558535123 podStartE2EDuration="3m7.558535123s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:49.553225723 +0000 UTC m=+251.654029222" watchObservedRunningTime="2026-03-17 09:13:49.558535123 +0000 UTC m=+251.659338622" Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.591276 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:49 crc kubenswrapper[4813]: E0317 09:13:49.591546 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:50.091535087 +0000 UTC m=+252.192338586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.643521 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8qg9n" podStartSLOduration=187.643503227 podStartE2EDuration="3m7.643503227s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:49.587755809 +0000 UTC m=+251.688559308" watchObservedRunningTime="2026-03-17 09:13:49.643503227 +0000 UTC m=+251.744306726" Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.692414 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:49 crc kubenswrapper[4813]: E0317 09:13:49.692760 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:50.192745954 +0000 UTC m=+252.293549453 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.773130 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ptq22"] Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.773378 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" podUID="55bb4fe2-94b0-4fdf-abb4-e39470fe2c30" containerName="controller-manager" containerID="cri-o://168060a45f4c11cf0e1cbc8021564701a55997c1a91b4a487bb37581d4c1f7d8" gracePeriod=30 Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.797290 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:49 crc kubenswrapper[4813]: E0317 09:13:49.797677 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:50.297664147 +0000 UTC m=+252.398467646 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.819005 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l"] Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.819211 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" podUID="768d1a31-0d78-4239-9d7d-260d11e4ad58" containerName="route-controller-manager" containerID="cri-o://120c9ea27cd7b3db8e036a33512fb311f5c7126c479216e881ff6ac734fdda55" gracePeriod=30 Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.910812 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:49 crc kubenswrapper[4813]: E0317 09:13:49.910955 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:50.410934374 +0000 UTC m=+252.511737873 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.911111 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:49 crc kubenswrapper[4813]: E0317 09:13:49.911554 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:50.411543355 +0000 UTC m=+252.512346854 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:49 crc kubenswrapper[4813]: I0317 09:13:49.998065 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-4rqzj" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.014379 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:50 crc kubenswrapper[4813]: E0317 09:13:50.014745 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:50.514729389 +0000 UTC m=+252.615532888 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.098994 4813 patch_prober.go:28] interesting pod/router-default-5444994796-m97th container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 09:13:50 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Mar 17 09:13:50 crc kubenswrapper[4813]: [+]process-running ok Mar 17 09:13:50 crc kubenswrapper[4813]: healthz check failed Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.099353 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-m97th" podUID="68399978-6f43-4458-a1e1-58d5984de8f5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.118222 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:50 crc kubenswrapper[4813]: E0317 09:13:50.118539 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:50.618526904 +0000 UTC m=+252.719330393 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.188226 4813 ???:1] "http: TLS handshake error from 192.168.126.11:52254: no serving certificate available for the kubelet" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.220932 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:50 crc kubenswrapper[4813]: E0317 09:13:50.221189 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:50.721161369 +0000 UTC m=+252.821964868 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.305698 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.324017 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-serving-cert\") pod \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\" (UID: \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\") " Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.324047 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ct9wj\" (UniqueName: \"kubernetes.io/projected/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-kube-api-access-ct9wj\") pod \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\" (UID: \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\") " Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.324094 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-config\") pod \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\" (UID: \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\") " Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.324109 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-client-ca\") pod \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\" (UID: \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\") " Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.324173 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-proxy-ca-bundles\") pod \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\" (UID: \"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30\") " Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.324377 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:50 crc kubenswrapper[4813]: E0317 09:13:50.324623 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:50.824612673 +0000 UTC m=+252.925416172 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.330447 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "55bb4fe2-94b0-4fdf-abb4-e39470fe2c30" (UID: "55bb4fe2-94b0-4fdf-abb4-e39470fe2c30"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.330945 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-client-ca" (OuterVolumeSpecName: "client-ca") pod "55bb4fe2-94b0-4fdf-abb4-e39470fe2c30" (UID: "55bb4fe2-94b0-4fdf-abb4-e39470fe2c30"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.337364 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-config" (OuterVolumeSpecName: "config") pod "55bb4fe2-94b0-4fdf-abb4-e39470fe2c30" (UID: "55bb4fe2-94b0-4fdf-abb4-e39470fe2c30"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.343026 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "55bb4fe2-94b0-4fdf-abb4-e39470fe2c30" (UID: "55bb4fe2-94b0-4fdf-abb4-e39470fe2c30"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.356425 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-kube-api-access-ct9wj" (OuterVolumeSpecName: "kube-api-access-ct9wj") pod "55bb4fe2-94b0-4fdf-abb4-e39470fe2c30" (UID: "55bb4fe2-94b0-4fdf-abb4-e39470fe2c30"). InnerVolumeSpecName "kube-api-access-ct9wj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.359728 4813 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-slqwx container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.42:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.359898 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx" podUID="9902dff6-4484-44ba-af31-877293debb91" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.42:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.374454 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.378578 4813 generic.go:334] "Generic (PLEG): container finished" podID="55bb4fe2-94b0-4fdf-abb4-e39470fe2c30" containerID="168060a45f4c11cf0e1cbc8021564701a55997c1a91b4a487bb37581d4c1f7d8" exitCode=0 Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.378641 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" event={"ID":"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30","Type":"ContainerDied","Data":"168060a45f4c11cf0e1cbc8021564701a55997c1a91b4a487bb37581d4c1f7d8"} Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.378665 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" event={"ID":"55bb4fe2-94b0-4fdf-abb4-e39470fe2c30","Type":"ContainerDied","Data":"56bc6b8828f391ce20802c03bbe556da08198fd0ad97378afbf331e3a28f1ad4"} Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.378679 4813 scope.go:117] "RemoveContainer" containerID="168060a45f4c11cf0e1cbc8021564701a55997c1a91b4a487bb37581d4c1f7d8" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.378761 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-ptq22" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.410914 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gczkp" event={"ID":"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b","Type":"ContainerStarted","Data":"e35e131200896acc61fd9d7952a0e72381177fecbab44a5c58ea8e535e43998e"} Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.411019 4813 scope.go:117] "RemoveContainer" containerID="168060a45f4c11cf0e1cbc8021564701a55997c1a91b4a487bb37581d4c1f7d8" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.434384 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.440215 4813 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 17 09:13:50 crc kubenswrapper[4813]: E0317 09:13:50.441923 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"168060a45f4c11cf0e1cbc8021564701a55997c1a91b4a487bb37581d4c1f7d8\": container with ID starting with 168060a45f4c11cf0e1cbc8021564701a55997c1a91b4a487bb37581d4c1f7d8 not found: ID does not exist" containerID="168060a45f4c11cf0e1cbc8021564701a55997c1a91b4a487bb37581d4c1f7d8" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.442061 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"168060a45f4c11cf0e1cbc8021564701a55997c1a91b4a487bb37581d4c1f7d8"} err="failed to get container status \"168060a45f4c11cf0e1cbc8021564701a55997c1a91b4a487bb37581d4c1f7d8\": rpc error: code = NotFound desc = could not find container \"168060a45f4c11cf0e1cbc8021564701a55997c1a91b4a487bb37581d4c1f7d8\": container with ID starting with 168060a45f4c11cf0e1cbc8021564701a55997c1a91b4a487bb37581d4c1f7d8 not found: ID does not exist" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.442859 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.442901 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ct9wj\" (UniqueName: \"kubernetes.io/projected/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-kube-api-access-ct9wj\") on node \"crc\" DevicePath \"\"" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.442912 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.442923 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.442935 4813 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 09:13:50 crc kubenswrapper[4813]: E0317 09:13:50.442989 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:50.942970733 +0000 UTC m=+253.043774232 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.468745 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ptq22"] Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.472184 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-ptq22"] Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.477927 4813 generic.go:334] "Generic (PLEG): container finished" podID="768d1a31-0d78-4239-9d7d-260d11e4ad58" containerID="120c9ea27cd7b3db8e036a33512fb311f5c7126c479216e881ff6ac734fdda55" exitCode=0 Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.478998 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.479343 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" event={"ID":"768d1a31-0d78-4239-9d7d-260d11e4ad58","Type":"ContainerDied","Data":"120c9ea27cd7b3db8e036a33512fb311f5c7126c479216e881ff6ac734fdda55"} Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.479365 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l" event={"ID":"768d1a31-0d78-4239-9d7d-260d11e4ad58","Type":"ContainerDied","Data":"6a5b92ae48f0bc497af13d4855416ea5532d65c8ef4a5339edb0bb79b8b84d4b"} Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.479389 4813 scope.go:117] "RemoveContainer" containerID="120c9ea27cd7b3db8e036a33512fb311f5c7126c479216e881ff6ac734fdda55" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.521569 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.530105 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-8qg9n" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.559097 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/768d1a31-0d78-4239-9d7d-260d11e4ad58-config\") pod \"768d1a31-0d78-4239-9d7d-260d11e4ad58\" (UID: \"768d1a31-0d78-4239-9d7d-260d11e4ad58\") " Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.559140 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/768d1a31-0d78-4239-9d7d-260d11e4ad58-serving-cert\") pod \"768d1a31-0d78-4239-9d7d-260d11e4ad58\" (UID: \"768d1a31-0d78-4239-9d7d-260d11e4ad58\") " Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.559164 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/768d1a31-0d78-4239-9d7d-260d11e4ad58-client-ca\") pod \"768d1a31-0d78-4239-9d7d-260d11e4ad58\" (UID: \"768d1a31-0d78-4239-9d7d-260d11e4ad58\") " Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.559204 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6m97\" (UniqueName: \"kubernetes.io/projected/768d1a31-0d78-4239-9d7d-260d11e4ad58-kube-api-access-d6m97\") pod \"768d1a31-0d78-4239-9d7d-260d11e4ad58\" (UID: \"768d1a31-0d78-4239-9d7d-260d11e4ad58\") " Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.560753 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:50 crc kubenswrapper[4813]: E0317 09:13:50.562027 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:51.062014207 +0000 UTC m=+253.162817706 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.570194 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/768d1a31-0d78-4239-9d7d-260d11e4ad58-client-ca" (OuterVolumeSpecName: "client-ca") pod "768d1a31-0d78-4239-9d7d-260d11e4ad58" (UID: "768d1a31-0d78-4239-9d7d-260d11e4ad58"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.570545 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/768d1a31-0d78-4239-9d7d-260d11e4ad58-config" (OuterVolumeSpecName: "config") pod "768d1a31-0d78-4239-9d7d-260d11e4ad58" (UID: "768d1a31-0d78-4239-9d7d-260d11e4ad58"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.574060 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/768d1a31-0d78-4239-9d7d-260d11e4ad58-kube-api-access-d6m97" (OuterVolumeSpecName: "kube-api-access-d6m97") pod "768d1a31-0d78-4239-9d7d-260d11e4ad58" (UID: "768d1a31-0d78-4239-9d7d-260d11e4ad58"). InnerVolumeSpecName "kube-api-access-d6m97". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.580435 4813 scope.go:117] "RemoveContainer" containerID="120c9ea27cd7b3db8e036a33512fb311f5c7126c479216e881ff6ac734fdda55" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.583736 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-slqwx" Mar 17 09:13:50 crc kubenswrapper[4813]: E0317 09:13:50.585547 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"120c9ea27cd7b3db8e036a33512fb311f5c7126c479216e881ff6ac734fdda55\": container with ID starting with 120c9ea27cd7b3db8e036a33512fb311f5c7126c479216e881ff6ac734fdda55 not found: ID does not exist" containerID="120c9ea27cd7b3db8e036a33512fb311f5c7126c479216e881ff6ac734fdda55" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.585803 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"120c9ea27cd7b3db8e036a33512fb311f5c7126c479216e881ff6ac734fdda55"} err="failed to get container status \"120c9ea27cd7b3db8e036a33512fb311f5c7126c479216e881ff6ac734fdda55\": rpc error: code = NotFound desc = could not find container \"120c9ea27cd7b3db8e036a33512fb311f5c7126c479216e881ff6ac734fdda55\": container with ID starting with 120c9ea27cd7b3db8e036a33512fb311f5c7126c479216e881ff6ac734fdda55 not found: ID does not exist" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.591981 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/768d1a31-0d78-4239-9d7d-260d11e4ad58-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "768d1a31-0d78-4239-9d7d-260d11e4ad58" (UID: "768d1a31-0d78-4239-9d7d-260d11e4ad58"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.600439 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6ff769775f-cz59s"] Mar 17 09:13:50 crc kubenswrapper[4813]: E0317 09:13:50.600710 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="768d1a31-0d78-4239-9d7d-260d11e4ad58" containerName="route-controller-manager" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.600726 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="768d1a31-0d78-4239-9d7d-260d11e4ad58" containerName="route-controller-manager" Mar 17 09:13:50 crc kubenswrapper[4813]: E0317 09:13:50.600742 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="55bb4fe2-94b0-4fdf-abb4-e39470fe2c30" containerName="controller-manager" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.600748 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="55bb4fe2-94b0-4fdf-abb4-e39470fe2c30" containerName="controller-manager" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.600882 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="768d1a31-0d78-4239-9d7d-260d11e4ad58" containerName="route-controller-manager" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.600903 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="55bb4fe2-94b0-4fdf-abb4-e39470fe2c30" containerName="controller-manager" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.601247 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.610366 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.610559 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.610648 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.610819 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.610983 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.611092 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.629242 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6ff769775f-cz59s"] Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.642992 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.654900 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr"] Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.660176 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.668198 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.668506 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a30c8ff-06c4-478a-bf04-21eb50da1d89-serving-cert\") pod \"controller-manager-6ff769775f-cz59s\" (UID: \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\") " pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.668572 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a30c8ff-06c4-478a-bf04-21eb50da1d89-config\") pod \"controller-manager-6ff769775f-cz59s\" (UID: \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\") " pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.668634 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a30c8ff-06c4-478a-bf04-21eb50da1d89-client-ca\") pod \"controller-manager-6ff769775f-cz59s\" (UID: \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\") " pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.668664 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blwtx\" (UniqueName: \"kubernetes.io/projected/1a30c8ff-06c4-478a-bf04-21eb50da1d89-kube-api-access-blwtx\") pod \"controller-manager-6ff769775f-cz59s\" (UID: \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\") " pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.668706 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1a30c8ff-06c4-478a-bf04-21eb50da1d89-proxy-ca-bundles\") pod \"controller-manager-6ff769775f-cz59s\" (UID: \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\") " pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" Mar 17 09:13:50 crc kubenswrapper[4813]: E0317 09:13:50.668761 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:51.168733362 +0000 UTC m=+253.269536861 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.668806 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.668922 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/768d1a31-0d78-4239-9d7d-260d11e4ad58-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.668933 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/768d1a31-0d78-4239-9d7d-260d11e4ad58-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.668941 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/768d1a31-0d78-4239-9d7d-260d11e4ad58-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.668950 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6m97\" (UniqueName: \"kubernetes.io/projected/768d1a31-0d78-4239-9d7d-260d11e4ad58-kube-api-access-d6m97\") on node \"crc\" DevicePath \"\"" Mar 17 09:13:50 crc kubenswrapper[4813]: E0317 09:13:50.669123 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:51.169109204 +0000 UTC m=+253.269912703 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.674476 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr"] Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.745128 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55bb4fe2-94b0-4fdf-abb4-e39470fe2c30" path="/var/lib/kubelet/pods/55bb4fe2-94b0-4fdf-abb4-e39470fe2c30/volumes" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.772074 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.772292 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdlsg\" (UniqueName: \"kubernetes.io/projected/e50bf93b-f84f-4050-b5cf-65733de8db5c-kube-api-access-jdlsg\") pod \"route-controller-manager-5885dd695f-qh6zr\" (UID: \"e50bf93b-f84f-4050-b5cf-65733de8db5c\") " pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.772324 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a30c8ff-06c4-478a-bf04-21eb50da1d89-client-ca\") pod \"controller-manager-6ff769775f-cz59s\" (UID: \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\") " pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.772358 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blwtx\" (UniqueName: \"kubernetes.io/projected/1a30c8ff-06c4-478a-bf04-21eb50da1d89-kube-api-access-blwtx\") pod \"controller-manager-6ff769775f-cz59s\" (UID: \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\") " pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.772383 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e50bf93b-f84f-4050-b5cf-65733de8db5c-serving-cert\") pod \"route-controller-manager-5885dd695f-qh6zr\" (UID: \"e50bf93b-f84f-4050-b5cf-65733de8db5c\") " pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.772396 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e50bf93b-f84f-4050-b5cf-65733de8db5c-client-ca\") pod \"route-controller-manager-5885dd695f-qh6zr\" (UID: \"e50bf93b-f84f-4050-b5cf-65733de8db5c\") " pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.772413 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1a30c8ff-06c4-478a-bf04-21eb50da1d89-proxy-ca-bundles\") pod \"controller-manager-6ff769775f-cz59s\" (UID: \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\") " pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.772463 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a30c8ff-06c4-478a-bf04-21eb50da1d89-serving-cert\") pod \"controller-manager-6ff769775f-cz59s\" (UID: \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\") " pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.772502 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a30c8ff-06c4-478a-bf04-21eb50da1d89-config\") pod \"controller-manager-6ff769775f-cz59s\" (UID: \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\") " pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.772520 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e50bf93b-f84f-4050-b5cf-65733de8db5c-config\") pod \"route-controller-manager-5885dd695f-qh6zr\" (UID: \"e50bf93b-f84f-4050-b5cf-65733de8db5c\") " pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.773837 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1a30c8ff-06c4-478a-bf04-21eb50da1d89-proxy-ca-bundles\") pod \"controller-manager-6ff769775f-cz59s\" (UID: \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\") " pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.774858 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a30c8ff-06c4-478a-bf04-21eb50da1d89-client-ca\") pod \"controller-manager-6ff769775f-cz59s\" (UID: \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\") " pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" Mar 17 09:13:50 crc kubenswrapper[4813]: E0317 09:13:50.774984 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:51.274967669 +0000 UTC m=+253.375771168 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.775004 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a30c8ff-06c4-478a-bf04-21eb50da1d89-config\") pod \"controller-manager-6ff769775f-cz59s\" (UID: \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\") " pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.777279 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a30c8ff-06c4-478a-bf04-21eb50da1d89-serving-cert\") pod \"controller-manager-6ff769775f-cz59s\" (UID: \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\") " pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.798297 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l"] Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.799181 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blwtx\" (UniqueName: \"kubernetes.io/projected/1a30c8ff-06c4-478a-bf04-21eb50da1d89-kube-api-access-blwtx\") pod \"controller-manager-6ff769775f-cz59s\" (UID: \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\") " pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.801397 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-6rn9l"] Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.874311 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.874379 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e50bf93b-f84f-4050-b5cf-65733de8db5c-config\") pod \"route-controller-manager-5885dd695f-qh6zr\" (UID: \"e50bf93b-f84f-4050-b5cf-65733de8db5c\") " pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.874411 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdlsg\" (UniqueName: \"kubernetes.io/projected/e50bf93b-f84f-4050-b5cf-65733de8db5c-kube-api-access-jdlsg\") pod \"route-controller-manager-5885dd695f-qh6zr\" (UID: \"e50bf93b-f84f-4050-b5cf-65733de8db5c\") " pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.874452 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e50bf93b-f84f-4050-b5cf-65733de8db5c-serving-cert\") pod \"route-controller-manager-5885dd695f-qh6zr\" (UID: \"e50bf93b-f84f-4050-b5cf-65733de8db5c\") " pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.874470 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e50bf93b-f84f-4050-b5cf-65733de8db5c-client-ca\") pod \"route-controller-manager-5885dd695f-qh6zr\" (UID: \"e50bf93b-f84f-4050-b5cf-65733de8db5c\") " pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" Mar 17 09:13:50 crc kubenswrapper[4813]: E0317 09:13:50.874710 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-17 09:13:51.374693415 +0000 UTC m=+253.475496914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dbm4b" (UID: "782b4936-2caa-44da-8716-c57924be6df3") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.875304 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e50bf93b-f84f-4050-b5cf-65733de8db5c-client-ca\") pod \"route-controller-manager-5885dd695f-qh6zr\" (UID: \"e50bf93b-f84f-4050-b5cf-65733de8db5c\") " pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.875974 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e50bf93b-f84f-4050-b5cf-65733de8db5c-config\") pod \"route-controller-manager-5885dd695f-qh6zr\" (UID: \"e50bf93b-f84f-4050-b5cf-65733de8db5c\") " pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.878181 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e50bf93b-f84f-4050-b5cf-65733de8db5c-serving-cert\") pod \"route-controller-manager-5885dd695f-qh6zr\" (UID: \"e50bf93b-f84f-4050-b5cf-65733de8db5c\") " pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.891238 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdlsg\" (UniqueName: \"kubernetes.io/projected/e50bf93b-f84f-4050-b5cf-65733de8db5c-kube-api-access-jdlsg\") pod \"route-controller-manager-5885dd695f-qh6zr\" (UID: \"e50bf93b-f84f-4050-b5cf-65733de8db5c\") " pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.965877 4813 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-17T09:13:50.44023801Z","Handler":null,"Name":""} Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.975310 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:50 crc kubenswrapper[4813]: E0317 09:13:50.975746 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-17 09:13:51.475722726 +0000 UTC m=+253.576526225 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.975948 4813 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.975978 4813 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 17 09:13:50 crc kubenswrapper[4813]: I0317 09:13:50.991006 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.007636 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.077072 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.083889 4813 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.083923 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.085270 4813 patch_prober.go:28] interesting pod/router-default-5444994796-m97th container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 09:13:51 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Mar 17 09:13:51 crc kubenswrapper[4813]: [+]process-running ok Mar 17 09:13:51 crc kubenswrapper[4813]: healthz check failed Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.085310 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-m97th" podUID="68399978-6f43-4458-a1e1-58d5984de8f5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.110149 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dbm4b\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.133652 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2skjk"] Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.134743 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2skjk" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.137056 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.159091 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2skjk"] Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.180613 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.180783 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6pmb\" (UniqueName: \"kubernetes.io/projected/65af6b35-41bd-4ad2-b8d4-0b653cb18dcf-kube-api-access-s6pmb\") pod \"certified-operators-2skjk\" (UID: \"65af6b35-41bd-4ad2-b8d4-0b653cb18dcf\") " pod="openshift-marketplace/certified-operators-2skjk" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.180850 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65af6b35-41bd-4ad2-b8d4-0b653cb18dcf-catalog-content\") pod \"certified-operators-2skjk\" (UID: \"65af6b35-41bd-4ad2-b8d4-0b653cb18dcf\") " pod="openshift-marketplace/certified-operators-2skjk" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.180871 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65af6b35-41bd-4ad2-b8d4-0b653cb18dcf-utilities\") pod \"certified-operators-2skjk\" (UID: \"65af6b35-41bd-4ad2-b8d4-0b653cb18dcf\") " pod="openshift-marketplace/certified-operators-2skjk" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.193540 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.282076 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6pmb\" (UniqueName: \"kubernetes.io/projected/65af6b35-41bd-4ad2-b8d4-0b653cb18dcf-kube-api-access-s6pmb\") pod \"certified-operators-2skjk\" (UID: \"65af6b35-41bd-4ad2-b8d4-0b653cb18dcf\") " pod="openshift-marketplace/certified-operators-2skjk" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.282215 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65af6b35-41bd-4ad2-b8d4-0b653cb18dcf-catalog-content\") pod \"certified-operators-2skjk\" (UID: \"65af6b35-41bd-4ad2-b8d4-0b653cb18dcf\") " pod="openshift-marketplace/certified-operators-2skjk" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.282242 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65af6b35-41bd-4ad2-b8d4-0b653cb18dcf-utilities\") pod \"certified-operators-2skjk\" (UID: \"65af6b35-41bd-4ad2-b8d4-0b653cb18dcf\") " pod="openshift-marketplace/certified-operators-2skjk" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.282913 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65af6b35-41bd-4ad2-b8d4-0b653cb18dcf-catalog-content\") pod \"certified-operators-2skjk\" (UID: \"65af6b35-41bd-4ad2-b8d4-0b653cb18dcf\") " pod="openshift-marketplace/certified-operators-2skjk" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.283953 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65af6b35-41bd-4ad2-b8d4-0b653cb18dcf-utilities\") pod \"certified-operators-2skjk\" (UID: \"65af6b35-41bd-4ad2-b8d4-0b653cb18dcf\") " pod="openshift-marketplace/certified-operators-2skjk" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.302030 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6pmb\" (UniqueName: \"kubernetes.io/projected/65af6b35-41bd-4ad2-b8d4-0b653cb18dcf-kube-api-access-s6pmb\") pod \"certified-operators-2skjk\" (UID: \"65af6b35-41bd-4ad2-b8d4-0b653cb18dcf\") " pod="openshift-marketplace/certified-operators-2skjk" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.322937 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.351379 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-lxrc7"] Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.357856 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lxrc7"] Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.359499 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lxrc7" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.363549 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.444723 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6ff769775f-cz59s"] Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.450580 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2skjk" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.486626 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b9e1de0-da0f-422e-a79b-4fb54d2e3acb-catalog-content\") pod \"community-operators-lxrc7\" (UID: \"1b9e1de0-da0f-422e-a79b-4fb54d2e3acb\") " pod="openshift-marketplace/community-operators-lxrc7" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.486683 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt5rs\" (UniqueName: \"kubernetes.io/projected/1b9e1de0-da0f-422e-a79b-4fb54d2e3acb-kube-api-access-tt5rs\") pod \"community-operators-lxrc7\" (UID: \"1b9e1de0-da0f-422e-a79b-4fb54d2e3acb\") " pod="openshift-marketplace/community-operators-lxrc7" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.486729 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b9e1de0-da0f-422e-a79b-4fb54d2e3acb-utilities\") pod \"community-operators-lxrc7\" (UID: \"1b9e1de0-da0f-422e-a79b-4fb54d2e3acb\") " pod="openshift-marketplace/community-operators-lxrc7" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.510334 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" event={"ID":"1a30c8ff-06c4-478a-bf04-21eb50da1d89","Type":"ContainerStarted","Data":"bee449c8ac9f58cd0f1cc43df8051ece39ec25f0129d0616913fc19ee6e6f89f"} Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.515697 4813 ???:1] "http: TLS handshake error from 192.168.126.11:52262: no serving certificate available for the kubelet" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.528042 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gczkp" event={"ID":"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b","Type":"ContainerStarted","Data":"3edca5c5a5766cc75115a491920a032733cb71a1883d674cd75cc5df1b41b283"} Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.528084 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr"] Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.528098 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-gczkp" event={"ID":"cfccf8af-20ed-4d82-96d7-b4b3c00cd24b","Type":"ContainerStarted","Data":"1acac8049b16283a1d6ae9e85f948d552c45005730e9f15a5fc88b6ff6e3c29a"} Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.540458 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5dpjf"] Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.542979 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5dpjf" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.546218 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-gczkp" podStartSLOduration=10.546203123 podStartE2EDuration="10.546203123s" podCreationTimestamp="2026-03-17 09:13:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:51.544281298 +0000 UTC m=+253.645084797" watchObservedRunningTime="2026-03-17 09:13:51.546203123 +0000 UTC m=+253.647006622" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.550641 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5dpjf"] Mar 17 09:13:51 crc kubenswrapper[4813]: W0317 09:13:51.558825 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode50bf93b_f84f_4050_b5cf_65733de8db5c.slice/crio-aa730b8da6c17173b497c1f2cd8704f3fe4819c17a6bf449ecd2e4f102aea7f2 WatchSource:0}: Error finding container aa730b8da6c17173b497c1f2cd8704f3fe4819c17a6bf449ecd2e4f102aea7f2: Status 404 returned error can't find the container with id aa730b8da6c17173b497c1f2cd8704f3fe4819c17a6bf449ecd2e4f102aea7f2 Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.582363 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dbm4b"] Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.590799 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b9e1de0-da0f-422e-a79b-4fb54d2e3acb-utilities\") pod \"community-operators-lxrc7\" (UID: \"1b9e1de0-da0f-422e-a79b-4fb54d2e3acb\") " pod="openshift-marketplace/community-operators-lxrc7" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.590880 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7d1b948-0f9a-4ba3-9f58-88590e13ba63-catalog-content\") pod \"certified-operators-5dpjf\" (UID: \"b7d1b948-0f9a-4ba3-9f58-88590e13ba63\") " pod="openshift-marketplace/certified-operators-5dpjf" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.590957 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b9e1de0-da0f-422e-a79b-4fb54d2e3acb-catalog-content\") pod \"community-operators-lxrc7\" (UID: \"1b9e1de0-da0f-422e-a79b-4fb54d2e3acb\") " pod="openshift-marketplace/community-operators-lxrc7" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.590984 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5vxln\" (UniqueName: \"kubernetes.io/projected/b7d1b948-0f9a-4ba3-9f58-88590e13ba63-kube-api-access-5vxln\") pod \"certified-operators-5dpjf\" (UID: \"b7d1b948-0f9a-4ba3-9f58-88590e13ba63\") " pod="openshift-marketplace/certified-operators-5dpjf" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.591063 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7d1b948-0f9a-4ba3-9f58-88590e13ba63-utilities\") pod \"certified-operators-5dpjf\" (UID: \"b7d1b948-0f9a-4ba3-9f58-88590e13ba63\") " pod="openshift-marketplace/certified-operators-5dpjf" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.591082 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt5rs\" (UniqueName: \"kubernetes.io/projected/1b9e1de0-da0f-422e-a79b-4fb54d2e3acb-kube-api-access-tt5rs\") pod \"community-operators-lxrc7\" (UID: \"1b9e1de0-da0f-422e-a79b-4fb54d2e3acb\") " pod="openshift-marketplace/community-operators-lxrc7" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.592785 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b9e1de0-da0f-422e-a79b-4fb54d2e3acb-utilities\") pod \"community-operators-lxrc7\" (UID: \"1b9e1de0-da0f-422e-a79b-4fb54d2e3acb\") " pod="openshift-marketplace/community-operators-lxrc7" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.593634 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b9e1de0-da0f-422e-a79b-4fb54d2e3acb-catalog-content\") pod \"community-operators-lxrc7\" (UID: \"1b9e1de0-da0f-422e-a79b-4fb54d2e3acb\") " pod="openshift-marketplace/community-operators-lxrc7" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.612440 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt5rs\" (UniqueName: \"kubernetes.io/projected/1b9e1de0-da0f-422e-a79b-4fb54d2e3acb-kube-api-access-tt5rs\") pod \"community-operators-lxrc7\" (UID: \"1b9e1de0-da0f-422e-a79b-4fb54d2e3acb\") " pod="openshift-marketplace/community-operators-lxrc7" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.682140 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lxrc7" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.695938 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5vxln\" (UniqueName: \"kubernetes.io/projected/b7d1b948-0f9a-4ba3-9f58-88590e13ba63-kube-api-access-5vxln\") pod \"certified-operators-5dpjf\" (UID: \"b7d1b948-0f9a-4ba3-9f58-88590e13ba63\") " pod="openshift-marketplace/certified-operators-5dpjf" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.696262 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.696285 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7d1b948-0f9a-4ba3-9f58-88590e13ba63-utilities\") pod \"certified-operators-5dpjf\" (UID: \"b7d1b948-0f9a-4ba3-9f58-88590e13ba63\") " pod="openshift-marketplace/certified-operators-5dpjf" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.696500 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.698475 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.698982 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7d1b948-0f9a-4ba3-9f58-88590e13ba63-utilities\") pod \"certified-operators-5dpjf\" (UID: \"b7d1b948-0f9a-4ba3-9f58-88590e13ba63\") " pod="openshift-marketplace/certified-operators-5dpjf" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.699056 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.699168 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7d1b948-0f9a-4ba3-9f58-88590e13ba63-catalog-content\") pod \"certified-operators-5dpjf\" (UID: \"b7d1b948-0f9a-4ba3-9f58-88590e13ba63\") " pod="openshift-marketplace/certified-operators-5dpjf" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.699356 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.699632 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7d1b948-0f9a-4ba3-9f58-88590e13ba63-catalog-content\") pod \"certified-operators-5dpjf\" (UID: \"b7d1b948-0f9a-4ba3-9f58-88590e13ba63\") " pod="openshift-marketplace/certified-operators-5dpjf" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.699660 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs\") pod \"network-metrics-daemon-l47ql\" (UID: \"27d4c184-5dd3-492a-b927-f7a7f31291ef\") " pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.699750 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.705335 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.705396 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.705335 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.707514 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.713952 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/27d4c184-5dd3-492a-b927-f7a7f31291ef-metrics-certs\") pod \"network-metrics-daemon-l47ql\" (UID: \"27d4c184-5dd3-492a-b927-f7a7f31291ef\") " pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.714067 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.716311 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2skjk"] Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.717438 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.717868 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.721554 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5vxln\" (UniqueName: \"kubernetes.io/projected/b7d1b948-0f9a-4ba3-9f58-88590e13ba63-kube-api-access-5vxln\") pod \"certified-operators-5dpjf\" (UID: \"b7d1b948-0f9a-4ba3-9f58-88590e13ba63\") " pod="openshift-marketplace/certified-operators-5dpjf" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.733272 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5lhdz"] Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.735666 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5lhdz" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.743740 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5lhdz"] Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.801672 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/404fc67e-83be-4be6-bf6e-4279c0543316-utilities\") pod \"community-operators-5lhdz\" (UID: \"404fc67e-83be-4be6-bf6e-4279c0543316\") " pod="openshift-marketplace/community-operators-5lhdz" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.802277 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/404fc67e-83be-4be6-bf6e-4279c0543316-catalog-content\") pod \"community-operators-5lhdz\" (UID: \"404fc67e-83be-4be6-bf6e-4279c0543316\") " pod="openshift-marketplace/community-operators-5lhdz" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.802334 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7vwj\" (UniqueName: \"kubernetes.io/projected/404fc67e-83be-4be6-bf6e-4279c0543316-kube-api-access-q7vwj\") pod \"community-operators-5lhdz\" (UID: \"404fc67e-83be-4be6-bf6e-4279c0543316\") " pod="openshift-marketplace/community-operators-5lhdz" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.856093 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.872380 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5dpjf" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.903427 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/404fc67e-83be-4be6-bf6e-4279c0543316-catalog-content\") pod \"community-operators-5lhdz\" (UID: \"404fc67e-83be-4be6-bf6e-4279c0543316\") " pod="openshift-marketplace/community-operators-5lhdz" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.903484 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7vwj\" (UniqueName: \"kubernetes.io/projected/404fc67e-83be-4be6-bf6e-4279c0543316-kube-api-access-q7vwj\") pod \"community-operators-5lhdz\" (UID: \"404fc67e-83be-4be6-bf6e-4279c0543316\") " pod="openshift-marketplace/community-operators-5lhdz" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.903538 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/404fc67e-83be-4be6-bf6e-4279c0543316-utilities\") pod \"community-operators-5lhdz\" (UID: \"404fc67e-83be-4be6-bf6e-4279c0543316\") " pod="openshift-marketplace/community-operators-5lhdz" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.904004 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/404fc67e-83be-4be6-bf6e-4279c0543316-utilities\") pod \"community-operators-5lhdz\" (UID: \"404fc67e-83be-4be6-bf6e-4279c0543316\") " pod="openshift-marketplace/community-operators-5lhdz" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.904276 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/404fc67e-83be-4be6-bf6e-4279c0543316-catalog-content\") pod \"community-operators-5lhdz\" (UID: \"404fc67e-83be-4be6-bf6e-4279c0543316\") " pod="openshift-marketplace/community-operators-5lhdz" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.941902 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7vwj\" (UniqueName: \"kubernetes.io/projected/404fc67e-83be-4be6-bf6e-4279c0543316-kube-api-access-q7vwj\") pod \"community-operators-5lhdz\" (UID: \"404fc67e-83be-4be6-bf6e-4279c0543316\") " pod="openshift-marketplace/community-operators-5lhdz" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.956862 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.975846 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.981320 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 17 09:13:51 crc kubenswrapper[4813]: I0317 09:13:51.982179 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-l47ql" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.011681 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-lxrc7"] Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.065920 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5lhdz" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.095898 4813 patch_prober.go:28] interesting pod/router-default-5444994796-m97th container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 09:13:52 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Mar 17 09:13:52 crc kubenswrapper[4813]: [+]process-running ok Mar 17 09:13:52 crc kubenswrapper[4813]: healthz check failed Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.095944 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-m97th" podUID="68399978-6f43-4458-a1e1-58d5984de8f5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.105442 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.106030 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.111873 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.112036 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.122288 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.207105 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/332581fa-6de2-41c0-9d20-f9d0b269c5df-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"332581fa-6de2-41c0-9d20-f9d0b269c5df\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.207189 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/332581fa-6de2-41c0-9d20-f9d0b269c5df-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"332581fa-6de2-41c0-9d20-f9d0b269c5df\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 17 09:13:52 crc kubenswrapper[4813]: W0317 09:13:52.302252 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-99d6160ff4eae499a93fb596cddbf52c6b40298cea82fb78681495a844cd2cda WatchSource:0}: Error finding container 99d6160ff4eae499a93fb596cddbf52c6b40298cea82fb78681495a844cd2cda: Status 404 returned error can't find the container with id 99d6160ff4eae499a93fb596cddbf52c6b40298cea82fb78681495a844cd2cda Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.310157 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/332581fa-6de2-41c0-9d20-f9d0b269c5df-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"332581fa-6de2-41c0-9d20-f9d0b269c5df\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.310308 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/332581fa-6de2-41c0-9d20-f9d0b269c5df-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"332581fa-6de2-41c0-9d20-f9d0b269c5df\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.311702 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/332581fa-6de2-41c0-9d20-f9d0b269c5df-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"332581fa-6de2-41c0-9d20-f9d0b269c5df\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.336710 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/332581fa-6de2-41c0-9d20-f9d0b269c5df-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"332581fa-6de2-41c0-9d20-f9d0b269c5df\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.457586 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.541263 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" event={"ID":"782b4936-2caa-44da-8716-c57924be6df3","Type":"ContainerStarted","Data":"be5066153cadae7bb93f3adaacd29a207801bea4da3722fc5f08063db6f705c8"} Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.541335 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.541346 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" event={"ID":"782b4936-2caa-44da-8716-c57924be6df3","Type":"ContainerStarted","Data":"0c096ac2c60f7bd51d4760d2758d8e014342c7765f19a533f26c9d79d8918b75"} Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.544384 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" event={"ID":"e50bf93b-f84f-4050-b5cf-65733de8db5c","Type":"ContainerStarted","Data":"fcd4566bc40b8b2584a7500e2ba1f341d08ae963f3c5c2af9fcb90ca3a290578"} Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.544417 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" event={"ID":"e50bf93b-f84f-4050-b5cf-65733de8db5c","Type":"ContainerStarted","Data":"aa730b8da6c17173b497c1f2cd8704f3fe4819c17a6bf449ecd2e4f102aea7f2"} Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.545078 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.550375 4813 generic.go:334] "Generic (PLEG): container finished" podID="b3992349-bc61-405e-be75-609cce2fad10" containerID="6abc6fdbecc5d29e40bf98aa15092c6a91d5d8667c098322c7987cf7b1896af4" exitCode=0 Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.550563 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7" event={"ID":"b3992349-bc61-405e-be75-609cce2fad10","Type":"ContainerDied","Data":"6abc6fdbecc5d29e40bf98aa15092c6a91d5d8667c098322c7987cf7b1896af4"} Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.553207 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.553619 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" event={"ID":"1a30c8ff-06c4-478a-bf04-21eb50da1d89","Type":"ContainerStarted","Data":"cbb747b8e38fa24891057d3d949205f36c5a00b26e7e343a7b0d4762b09840fb"} Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.553951 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.555255 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"99d6160ff4eae499a93fb596cddbf52c6b40298cea82fb78681495a844cd2cda"} Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.556733 4813 generic.go:334] "Generic (PLEG): container finished" podID="1b9e1de0-da0f-422e-a79b-4fb54d2e3acb" containerID="22466df678fe5db67f486cbd971560fd45a8773a011b67e24d9fcbfe70ea5437" exitCode=0 Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.556810 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxrc7" event={"ID":"1b9e1de0-da0f-422e-a79b-4fb54d2e3acb","Type":"ContainerDied","Data":"22466df678fe5db67f486cbd971560fd45a8773a011b67e24d9fcbfe70ea5437"} Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.556829 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxrc7" event={"ID":"1b9e1de0-da0f-422e-a79b-4fb54d2e3acb","Type":"ContainerStarted","Data":"8de6c471cf2943414655249287ef4a24d83e123ebcca15674b298e837c07d55d"} Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.559784 4813 generic.go:334] "Generic (PLEG): container finished" podID="65af6b35-41bd-4ad2-b8d4-0b653cb18dcf" containerID="9711ceab7b56277dc92c943aff876676bcfc90a4859e2785dbaecb6b9be8644f" exitCode=0 Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.560918 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2skjk" event={"ID":"65af6b35-41bd-4ad2-b8d4-0b653cb18dcf","Type":"ContainerDied","Data":"9711ceab7b56277dc92c943aff876676bcfc90a4859e2785dbaecb6b9be8644f"} Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.561012 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2skjk" event={"ID":"65af6b35-41bd-4ad2-b8d4-0b653cb18dcf","Type":"ContainerStarted","Data":"9d53eb7a90a322ac2991f4e21796bfce267b374b8627c60a7de116768b7f550d"} Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.581189 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.599790 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" podStartSLOduration=190.599764583 podStartE2EDuration="3m10.599764583s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:52.569638477 +0000 UTC m=+254.670441976" watchObservedRunningTime="2026-03-17 09:13:52.599764583 +0000 UTC m=+254.700568082" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.606293 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.607440 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.610400 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" podStartSLOduration=2.610377684 podStartE2EDuration="2.610377684s" podCreationTimestamp="2026-03-17 09:13:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:52.598498989 +0000 UTC m=+254.699302488" watchObservedRunningTime="2026-03-17 09:13:52.610377684 +0000 UTC m=+254.711181183" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.613612 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.613975 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.614863 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.618305 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5dpjf"] Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.621385 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.631434 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-kns96" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.702983 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" podStartSLOduration=2.702953937 podStartE2EDuration="2.702953937s" podCreationTimestamp="2026-03-17 09:13:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:52.675490262 +0000 UTC m=+254.776293761" watchObservedRunningTime="2026-03-17 09:13:52.702953937 +0000 UTC m=+254.803757436" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.773015 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="768d1a31-0d78-4239-9d7d-260d11e4ad58" path="/var/lib/kubelet/pods/768d1a31-0d78-4239-9d7d-260d11e4ad58/volumes" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.773800 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.785981 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.817549 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8f167ed7-d30d-4757-b673-3766ac2a113e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"8f167ed7-d30d-4757-b673-3766ac2a113e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.817726 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8f167ed7-d30d-4757-b673-3766ac2a113e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"8f167ed7-d30d-4757-b673-3766ac2a113e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.867333 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5lhdz"] Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.889147 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-l47ql"] Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.918448 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8f167ed7-d30d-4757-b673-3766ac2a113e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"8f167ed7-d30d-4757-b673-3766ac2a113e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.918520 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8f167ed7-d30d-4757-b673-3766ac2a113e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"8f167ed7-d30d-4757-b673-3766ac2a113e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.918584 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8f167ed7-d30d-4757-b673-3766ac2a113e-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"8f167ed7-d30d-4757-b673-3766ac2a113e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 17 09:13:52 crc kubenswrapper[4813]: I0317 09:13:52.946644 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8f167ed7-d30d-4757-b673-3766ac2a113e-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"8f167ed7-d30d-4757-b673-3766ac2a113e\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.085324 4813 patch_prober.go:28] interesting pod/router-default-5444994796-m97th container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 09:13:53 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Mar 17 09:13:53 crc kubenswrapper[4813]: [+]process-running ok Mar 17 09:13:53 crc kubenswrapper[4813]: healthz check failed Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.085372 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-m97th" podUID="68399978-6f43-4458-a1e1-58d5984de8f5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.146206 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-js422"] Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.147305 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-js422" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.159679 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.164384 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-js422"] Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.229928 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.325944 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12cf87e0-5d1a-4598-bd14-e3ef92a56dc1-utilities\") pod \"redhat-marketplace-js422\" (UID: \"12cf87e0-5d1a-4598-bd14-e3ef92a56dc1\") " pod="openshift-marketplace/redhat-marketplace-js422" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.325984 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wzmfj\" (UniqueName: \"kubernetes.io/projected/12cf87e0-5d1a-4598-bd14-e3ef92a56dc1-kube-api-access-wzmfj\") pod \"redhat-marketplace-js422\" (UID: \"12cf87e0-5d1a-4598-bd14-e3ef92a56dc1\") " pod="openshift-marketplace/redhat-marketplace-js422" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.326069 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12cf87e0-5d1a-4598-bd14-e3ef92a56dc1-catalog-content\") pod \"redhat-marketplace-js422\" (UID: \"12cf87e0-5d1a-4598-bd14-e3ef92a56dc1\") " pod="openshift-marketplace/redhat-marketplace-js422" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.427204 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wzmfj\" (UniqueName: \"kubernetes.io/projected/12cf87e0-5d1a-4598-bd14-e3ef92a56dc1-kube-api-access-wzmfj\") pod \"redhat-marketplace-js422\" (UID: \"12cf87e0-5d1a-4598-bd14-e3ef92a56dc1\") " pod="openshift-marketplace/redhat-marketplace-js422" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.427843 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12cf87e0-5d1a-4598-bd14-e3ef92a56dc1-catalog-content\") pod \"redhat-marketplace-js422\" (UID: \"12cf87e0-5d1a-4598-bd14-e3ef92a56dc1\") " pod="openshift-marketplace/redhat-marketplace-js422" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.427884 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12cf87e0-5d1a-4598-bd14-e3ef92a56dc1-utilities\") pod \"redhat-marketplace-js422\" (UID: \"12cf87e0-5d1a-4598-bd14-e3ef92a56dc1\") " pod="openshift-marketplace/redhat-marketplace-js422" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.428472 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12cf87e0-5d1a-4598-bd14-e3ef92a56dc1-utilities\") pod \"redhat-marketplace-js422\" (UID: \"12cf87e0-5d1a-4598-bd14-e3ef92a56dc1\") " pod="openshift-marketplace/redhat-marketplace-js422" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.428793 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12cf87e0-5d1a-4598-bd14-e3ef92a56dc1-catalog-content\") pod \"redhat-marketplace-js422\" (UID: \"12cf87e0-5d1a-4598-bd14-e3ef92a56dc1\") " pod="openshift-marketplace/redhat-marketplace-js422" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.443204 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.451254 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wzmfj\" (UniqueName: \"kubernetes.io/projected/12cf87e0-5d1a-4598-bd14-e3ef92a56dc1-kube-api-access-wzmfj\") pod \"redhat-marketplace-js422\" (UID: \"12cf87e0-5d1a-4598-bd14-e3ef92a56dc1\") " pod="openshift-marketplace/redhat-marketplace-js422" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.484241 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-js422" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.536092 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-x95rg"] Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.537194 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x95rg" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.550257 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x95rg"] Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.587759 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-l47ql" event={"ID":"27d4c184-5dd3-492a-b927-f7a7f31291ef","Type":"ContainerStarted","Data":"a917fcde01ec04dfba73a5ee2ba09f863b30d0dbefd8f785369d136cb4260f5f"} Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.587813 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-l47ql" event={"ID":"27d4c184-5dd3-492a-b927-f7a7f31291ef","Type":"ContainerStarted","Data":"c79b1d40586f402741f0d5263195454151ac5798d764f7f5d9abea29f50536c2"} Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.588261 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.592744 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"82401731de4aa09cb579bd5cf559277d49a65c409e920cd3258810ee5f6017d0"} Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.592787 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"8d45a3fe7fd5479d9bed0683caf152e80fd855fa6df69f169cfcd5b26bae5b5f"} Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.592976 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.593654 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-h87sm" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.613281 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"603c48976966028ce3f54840e10f01bcc515a981a0760a7d4f362c4151b76664"} Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.614609 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"8f167ed7-d30d-4757-b673-3766ac2a113e","Type":"ContainerStarted","Data":"ed1dbc491bd6599e329285675094273014210d797f3709a3e302215e91b6a874"} Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.615973 4813 generic.go:334] "Generic (PLEG): container finished" podID="404fc67e-83be-4be6-bf6e-4279c0543316" containerID="906a9701eab35212e064aab915a49fbd3f1bf2006d8079b721076b375f00ff18" exitCode=0 Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.616024 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5lhdz" event={"ID":"404fc67e-83be-4be6-bf6e-4279c0543316","Type":"ContainerDied","Data":"906a9701eab35212e064aab915a49fbd3f1bf2006d8079b721076b375f00ff18"} Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.616043 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5lhdz" event={"ID":"404fc67e-83be-4be6-bf6e-4279c0543316","Type":"ContainerStarted","Data":"b60bb60cfafc59976f24900693c81a01dc2dbf5000f8873a6a74143b1ac0ad0b"} Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.620459 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"332581fa-6de2-41c0-9d20-f9d0b269c5df","Type":"ContainerStarted","Data":"88d160517633d9cac0c5b8313a1e102ffd635d90c75c9af060627fb350daff42"} Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.620489 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"332581fa-6de2-41c0-9d20-f9d0b269c5df","Type":"ContainerStarted","Data":"c9dac6fdab420e1fb6834bbd5d6f4effe83ba541c62f6ab096e6a30c8327930d"} Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.631960 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"baf55ed3d1af165dfb85b05f3afcb028022df2c2c1b85437bfd48c04ee802d18"} Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.632008 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"68f93b19467cad46639b6fc88b11bf04305ed86dc6ecb447b80e26a1253ab432"} Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.637154 4813 generic.go:334] "Generic (PLEG): container finished" podID="b7d1b948-0f9a-4ba3-9f58-88590e13ba63" containerID="a6952169c00766088a53864389b265cec117256844467b5537f5d7d067ec0459" exitCode=0 Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.637299 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5dpjf" event={"ID":"b7d1b948-0f9a-4ba3-9f58-88590e13ba63","Type":"ContainerDied","Data":"a6952169c00766088a53864389b265cec117256844467b5537f5d7d067ec0459"} Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.637362 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5dpjf" event={"ID":"b7d1b948-0f9a-4ba3-9f58-88590e13ba63","Type":"ContainerStarted","Data":"d93323d902b018edef2b9ab9bdf52561a8da84a2a7057d1376b6383b00a51846"} Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.735016 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=1.734994943 podStartE2EDuration="1.734994943s" podCreationTimestamp="2026-03-17 09:13:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:53.72198387 +0000 UTC m=+255.822787369" watchObservedRunningTime="2026-03-17 09:13:53.734994943 +0000 UTC m=+255.835798442" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.737376 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6z7ks\" (UniqueName: \"kubernetes.io/projected/4ca1cd05-e8fc-47b7-904b-a0a544f654b0-kube-api-access-6z7ks\") pod \"redhat-marketplace-x95rg\" (UID: \"4ca1cd05-e8fc-47b7-904b-a0a544f654b0\") " pod="openshift-marketplace/redhat-marketplace-x95rg" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.737501 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ca1cd05-e8fc-47b7-904b-a0a544f654b0-catalog-content\") pod \"redhat-marketplace-x95rg\" (UID: \"4ca1cd05-e8fc-47b7-904b-a0a544f654b0\") " pod="openshift-marketplace/redhat-marketplace-x95rg" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.737587 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ca1cd05-e8fc-47b7-904b-a0a544f654b0-utilities\") pod \"redhat-marketplace-x95rg\" (UID: \"4ca1cd05-e8fc-47b7-904b-a0a544f654b0\") " pod="openshift-marketplace/redhat-marketplace-x95rg" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.838362 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ca1cd05-e8fc-47b7-904b-a0a544f654b0-catalog-content\") pod \"redhat-marketplace-x95rg\" (UID: \"4ca1cd05-e8fc-47b7-904b-a0a544f654b0\") " pod="openshift-marketplace/redhat-marketplace-x95rg" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.838520 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ca1cd05-e8fc-47b7-904b-a0a544f654b0-utilities\") pod \"redhat-marketplace-x95rg\" (UID: \"4ca1cd05-e8fc-47b7-904b-a0a544f654b0\") " pod="openshift-marketplace/redhat-marketplace-x95rg" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.838657 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6z7ks\" (UniqueName: \"kubernetes.io/projected/4ca1cd05-e8fc-47b7-904b-a0a544f654b0-kube-api-access-6z7ks\") pod \"redhat-marketplace-x95rg\" (UID: \"4ca1cd05-e8fc-47b7-904b-a0a544f654b0\") " pod="openshift-marketplace/redhat-marketplace-x95rg" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.844054 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ca1cd05-e8fc-47b7-904b-a0a544f654b0-catalog-content\") pod \"redhat-marketplace-x95rg\" (UID: \"4ca1cd05-e8fc-47b7-904b-a0a544f654b0\") " pod="openshift-marketplace/redhat-marketplace-x95rg" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.853811 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ca1cd05-e8fc-47b7-904b-a0a544f654b0-utilities\") pod \"redhat-marketplace-x95rg\" (UID: \"4ca1cd05-e8fc-47b7-904b-a0a544f654b0\") " pod="openshift-marketplace/redhat-marketplace-x95rg" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.867483 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6z7ks\" (UniqueName: \"kubernetes.io/projected/4ca1cd05-e8fc-47b7-904b-a0a544f654b0-kube-api-access-6z7ks\") pod \"redhat-marketplace-x95rg\" (UID: \"4ca1cd05-e8fc-47b7-904b-a0a544f654b0\") " pod="openshift-marketplace/redhat-marketplace-x95rg" Mar 17 09:13:53 crc kubenswrapper[4813]: I0317 09:13:53.885132 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x95rg" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.082037 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-m97th" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.084800 4813 patch_prober.go:28] interesting pod/router-default-5444994796-m97th container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 09:13:54 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Mar 17 09:13:54 crc kubenswrapper[4813]: [+]process-running ok Mar 17 09:13:54 crc kubenswrapper[4813]: healthz check failed Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.084839 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-m97th" podUID="68399978-6f43-4458-a1e1-58d5984de8f5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.106681 4813 ???:1] "http: TLS handshake error from 192.168.126.11:52274: no serving certificate available for the kubelet" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.154143 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.247395 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n26g7\" (UniqueName: \"kubernetes.io/projected/b3992349-bc61-405e-be75-609cce2fad10-kube-api-access-n26g7\") pod \"b3992349-bc61-405e-be75-609cce2fad10\" (UID: \"b3992349-bc61-405e-be75-609cce2fad10\") " Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.247438 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3992349-bc61-405e-be75-609cce2fad10-secret-volume\") pod \"b3992349-bc61-405e-be75-609cce2fad10\" (UID: \"b3992349-bc61-405e-be75-609cce2fad10\") " Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.247461 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3992349-bc61-405e-be75-609cce2fad10-config-volume\") pod \"b3992349-bc61-405e-be75-609cce2fad10\" (UID: \"b3992349-bc61-405e-be75-609cce2fad10\") " Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.248148 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b3992349-bc61-405e-be75-609cce2fad10-config-volume" (OuterVolumeSpecName: "config-volume") pod "b3992349-bc61-405e-be75-609cce2fad10" (UID: "b3992349-bc61-405e-be75-609cce2fad10"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.248555 4813 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3992349-bc61-405e-be75-609cce2fad10-config-volume\") on node \"crc\" DevicePath \"\"" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.254746 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b3992349-bc61-405e-be75-609cce2fad10-kube-api-access-n26g7" (OuterVolumeSpecName: "kube-api-access-n26g7") pod "b3992349-bc61-405e-be75-609cce2fad10" (UID: "b3992349-bc61-405e-be75-609cce2fad10"). InnerVolumeSpecName "kube-api-access-n26g7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.255535 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b3992349-bc61-405e-be75-609cce2fad10-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b3992349-bc61-405e-be75-609cce2fad10" (UID: "b3992349-bc61-405e-be75-609cce2fad10"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.335676 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-js422"] Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.343395 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6gvsx"] Mar 17 09:13:54 crc kubenswrapper[4813]: E0317 09:13:54.343727 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b3992349-bc61-405e-be75-609cce2fad10" containerName="collect-profiles" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.343762 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b3992349-bc61-405e-be75-609cce2fad10" containerName="collect-profiles" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.343915 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b3992349-bc61-405e-be75-609cce2fad10" containerName="collect-profiles" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.344680 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6gvsx" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.346934 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.353423 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n26g7\" (UniqueName: \"kubernetes.io/projected/b3992349-bc61-405e-be75-609cce2fad10-kube-api-access-n26g7\") on node \"crc\" DevicePath \"\"" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.353439 4813 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b3992349-bc61-405e-be75-609cce2fad10-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.356039 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6gvsx"] Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.363774 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-xb4k9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.363814 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-xb4k9" podUID="3aa0fc32-ed11-482b-9a6d-fcd59544bf8d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.364285 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-xb4k9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.364464 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xb4k9" podUID="3aa0fc32-ed11-482b-9a6d-fcd59544bf8d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.417403 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.417446 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.418662 4813 patch_prober.go:28] interesting pod/console-f9d7485db-2w9vb container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" start-of-body= Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.418717 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-2w9vb" podUID="f7603094-bb1d-4d77-b463-2270003b2805" containerName="console" probeResult="failure" output="Get \"https://10.217.0.17:8443/health\": dial tcp 10.217.0.17:8443: connect: connection refused" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.454634 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b3a014a-3574-4c2b-8331-df326d47f240-utilities\") pod \"redhat-operators-6gvsx\" (UID: \"6b3a014a-3574-4c2b-8331-df326d47f240\") " pod="openshift-marketplace/redhat-operators-6gvsx" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.454686 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b3a014a-3574-4c2b-8331-df326d47f240-catalog-content\") pod \"redhat-operators-6gvsx\" (UID: \"6b3a014a-3574-4c2b-8331-df326d47f240\") " pod="openshift-marketplace/redhat-operators-6gvsx" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.454826 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chhpl\" (UniqueName: \"kubernetes.io/projected/6b3a014a-3574-4c2b-8331-df326d47f240-kube-api-access-chhpl\") pod \"redhat-operators-6gvsx\" (UID: \"6b3a014a-3574-4c2b-8331-df326d47f240\") " pod="openshift-marketplace/redhat-operators-6gvsx" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.482488 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-x95rg"] Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.555884 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chhpl\" (UniqueName: \"kubernetes.io/projected/6b3a014a-3574-4c2b-8331-df326d47f240-kube-api-access-chhpl\") pod \"redhat-operators-6gvsx\" (UID: \"6b3a014a-3574-4c2b-8331-df326d47f240\") " pod="openshift-marketplace/redhat-operators-6gvsx" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.555958 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b3a014a-3574-4c2b-8331-df326d47f240-utilities\") pod \"redhat-operators-6gvsx\" (UID: \"6b3a014a-3574-4c2b-8331-df326d47f240\") " pod="openshift-marketplace/redhat-operators-6gvsx" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.555980 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b3a014a-3574-4c2b-8331-df326d47f240-catalog-content\") pod \"redhat-operators-6gvsx\" (UID: \"6b3a014a-3574-4c2b-8331-df326d47f240\") " pod="openshift-marketplace/redhat-operators-6gvsx" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.559185 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b3a014a-3574-4c2b-8331-df326d47f240-utilities\") pod \"redhat-operators-6gvsx\" (UID: \"6b3a014a-3574-4c2b-8331-df326d47f240\") " pod="openshift-marketplace/redhat-operators-6gvsx" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.559793 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b3a014a-3574-4c2b-8331-df326d47f240-catalog-content\") pod \"redhat-operators-6gvsx\" (UID: \"6b3a014a-3574-4c2b-8331-df326d47f240\") " pod="openshift-marketplace/redhat-operators-6gvsx" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.584958 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chhpl\" (UniqueName: \"kubernetes.io/projected/6b3a014a-3574-4c2b-8331-df326d47f240-kube-api-access-chhpl\") pod \"redhat-operators-6gvsx\" (UID: \"6b3a014a-3574-4c2b-8331-df326d47f240\") " pod="openshift-marketplace/redhat-operators-6gvsx" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.666148 4813 generic.go:334] "Generic (PLEG): container finished" podID="12cf87e0-5d1a-4598-bd14-e3ef92a56dc1" containerID="20728e6f4ee982c6741a69a4a1b812c4e55d6d59c4b07372616ddaf75192f20b" exitCode=0 Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.666209 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-js422" event={"ID":"12cf87e0-5d1a-4598-bd14-e3ef92a56dc1","Type":"ContainerDied","Data":"20728e6f4ee982c6741a69a4a1b812c4e55d6d59c4b07372616ddaf75192f20b"} Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.666290 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-js422" event={"ID":"12cf87e0-5d1a-4598-bd14-e3ef92a56dc1","Type":"ContainerStarted","Data":"b3f529df3e9fb0355f893cedef76587326c48494900c6d1e92d79a85d5638f2d"} Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.692097 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6gvsx" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.720906 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x95rg" event={"ID":"4ca1cd05-e8fc-47b7-904b-a0a544f654b0","Type":"ContainerStarted","Data":"d464b169c60197b9c240cfa168fe7b6ef35f19e0d369fca2260676bae19969cc"} Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.723406 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7" event={"ID":"b3992349-bc61-405e-be75-609cce2fad10","Type":"ContainerDied","Data":"44248ad0ada19b5a9f1ab05fa4dc5f199ea2c9e424f5d54e750f54345d728673"} Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.723467 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44248ad0ada19b5a9f1ab05fa4dc5f199ea2c9e424f5d54e750f54345d728673" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.723554 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.774084 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-l47ql" event={"ID":"27d4c184-5dd3-492a-b927-f7a7f31291ef","Type":"ContainerStarted","Data":"a60bebf89c0bb14d5d6814f58abaf387138957379d642f13aa08f8d1f82202f8"} Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.774152 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-bf7m2"] Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.775664 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bf7m2"] Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.775797 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bf7m2" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.787124 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"8f167ed7-d30d-4757-b673-3766ac2a113e","Type":"ContainerStarted","Data":"8b6ff5973a2d7d880ffb4097bdbd7f7d1dbadbc26087415b8e3505aff9e0d647"} Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.794643 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-l47ql" podStartSLOduration=192.794619519 podStartE2EDuration="3m12.794619519s" podCreationTimestamp="2026-03-17 09:10:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:54.791590565 +0000 UTC m=+256.892394064" watchObservedRunningTime="2026-03-17 09:13:54.794619519 +0000 UTC m=+256.895423018" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.808204 4813 generic.go:334] "Generic (PLEG): container finished" podID="332581fa-6de2-41c0-9d20-f9d0b269c5df" containerID="88d160517633d9cac0c5b8313a1e102ffd635d90c75c9af060627fb350daff42" exitCode=0 Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.809495 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"332581fa-6de2-41c0-9d20-f9d0b269c5df","Type":"ContainerDied","Data":"88d160517633d9cac0c5b8313a1e102ffd635d90c75c9af060627fb350daff42"} Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.813940 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.813917416 podStartE2EDuration="2.813917416s" podCreationTimestamp="2026-03-17 09:13:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:13:54.810559941 +0000 UTC m=+256.911363440" watchObservedRunningTime="2026-03-17 09:13:54.813917416 +0000 UTC m=+256.914720905" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.858990 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blzbh\" (UniqueName: \"kubernetes.io/projected/e49eaa7a-38bf-4996-ad34-64959cbccda2-kube-api-access-blzbh\") pod \"redhat-operators-bf7m2\" (UID: \"e49eaa7a-38bf-4996-ad34-64959cbccda2\") " pod="openshift-marketplace/redhat-operators-bf7m2" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.859103 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e49eaa7a-38bf-4996-ad34-64959cbccda2-catalog-content\") pod \"redhat-operators-bf7m2\" (UID: \"e49eaa7a-38bf-4996-ad34-64959cbccda2\") " pod="openshift-marketplace/redhat-operators-bf7m2" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.859226 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e49eaa7a-38bf-4996-ad34-64959cbccda2-utilities\") pod \"redhat-operators-bf7m2\" (UID: \"e49eaa7a-38bf-4996-ad34-64959cbccda2\") " pod="openshift-marketplace/redhat-operators-bf7m2" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.960082 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e49eaa7a-38bf-4996-ad34-64959cbccda2-catalog-content\") pod \"redhat-operators-bf7m2\" (UID: \"e49eaa7a-38bf-4996-ad34-64959cbccda2\") " pod="openshift-marketplace/redhat-operators-bf7m2" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.960285 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e49eaa7a-38bf-4996-ad34-64959cbccda2-utilities\") pod \"redhat-operators-bf7m2\" (UID: \"e49eaa7a-38bf-4996-ad34-64959cbccda2\") " pod="openshift-marketplace/redhat-operators-bf7m2" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.960312 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blzbh\" (UniqueName: \"kubernetes.io/projected/e49eaa7a-38bf-4996-ad34-64959cbccda2-kube-api-access-blzbh\") pod \"redhat-operators-bf7m2\" (UID: \"e49eaa7a-38bf-4996-ad34-64959cbccda2\") " pod="openshift-marketplace/redhat-operators-bf7m2" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.962765 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e49eaa7a-38bf-4996-ad34-64959cbccda2-catalog-content\") pod \"redhat-operators-bf7m2\" (UID: \"e49eaa7a-38bf-4996-ad34-64959cbccda2\") " pod="openshift-marketplace/redhat-operators-bf7m2" Mar 17 09:13:54 crc kubenswrapper[4813]: I0317 09:13:54.964098 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e49eaa7a-38bf-4996-ad34-64959cbccda2-utilities\") pod \"redhat-operators-bf7m2\" (UID: \"e49eaa7a-38bf-4996-ad34-64959cbccda2\") " pod="openshift-marketplace/redhat-operators-bf7m2" Mar 17 09:13:55 crc kubenswrapper[4813]: I0317 09:13:55.008556 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blzbh\" (UniqueName: \"kubernetes.io/projected/e49eaa7a-38bf-4996-ad34-64959cbccda2-kube-api-access-blzbh\") pod \"redhat-operators-bf7m2\" (UID: \"e49eaa7a-38bf-4996-ad34-64959cbccda2\") " pod="openshift-marketplace/redhat-operators-bf7m2" Mar 17 09:13:55 crc kubenswrapper[4813]: I0317 09:13:55.089733 4813 patch_prober.go:28] interesting pod/router-default-5444994796-m97th container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 17 09:13:55 crc kubenswrapper[4813]: [-]has-synced failed: reason withheld Mar 17 09:13:55 crc kubenswrapper[4813]: [+]process-running ok Mar 17 09:13:55 crc kubenswrapper[4813]: healthz check failed Mar 17 09:13:55 crc kubenswrapper[4813]: I0317 09:13:55.089818 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-m97th" podUID="68399978-6f43-4458-a1e1-58d5984de8f5" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 17 09:13:55 crc kubenswrapper[4813]: I0317 09:13:55.108838 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bf7m2" Mar 17 09:13:55 crc kubenswrapper[4813]: I0317 09:13:55.166903 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6gvsx"] Mar 17 09:13:55 crc kubenswrapper[4813]: I0317 09:13:55.833360 4813 generic.go:334] "Generic (PLEG): container finished" podID="8f167ed7-d30d-4757-b673-3766ac2a113e" containerID="8b6ff5973a2d7d880ffb4097bdbd7f7d1dbadbc26087415b8e3505aff9e0d647" exitCode=0 Mar 17 09:13:55 crc kubenswrapper[4813]: I0317 09:13:55.833454 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"8f167ed7-d30d-4757-b673-3766ac2a113e","Type":"ContainerDied","Data":"8b6ff5973a2d7d880ffb4097bdbd7f7d1dbadbc26087415b8e3505aff9e0d647"} Mar 17 09:13:55 crc kubenswrapper[4813]: I0317 09:13:55.849452 4813 generic.go:334] "Generic (PLEG): container finished" podID="4ca1cd05-e8fc-47b7-904b-a0a544f654b0" containerID="41421cebc179c392b9464e16fdc9c15c1c1e7aef5e6ba069628a436d12d3ed22" exitCode=0 Mar 17 09:13:55 crc kubenswrapper[4813]: I0317 09:13:55.849640 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x95rg" event={"ID":"4ca1cd05-e8fc-47b7-904b-a0a544f654b0","Type":"ContainerDied","Data":"41421cebc179c392b9464e16fdc9c15c1c1e7aef5e6ba069628a436d12d3ed22"} Mar 17 09:13:56 crc kubenswrapper[4813]: I0317 09:13:56.091745 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-m97th" Mar 17 09:13:56 crc kubenswrapper[4813]: I0317 09:13:56.098265 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-m97th" Mar 17 09:13:57 crc kubenswrapper[4813]: I0317 09:13:57.077930 4813 ???:1] "http: TLS handshake error from 192.168.126.11:46828: no serving certificate available for the kubelet" Mar 17 09:13:57 crc kubenswrapper[4813]: I0317 09:13:57.872383 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-665b6dd947-dpghh_845ee13b-7ed9-4910-958b-5269812237f0/cluster-samples-operator/0.log" Mar 17 09:13:57 crc kubenswrapper[4813]: I0317 09:13:57.872800 4813 generic.go:334] "Generic (PLEG): container finished" podID="845ee13b-7ed9-4910-958b-5269812237f0" containerID="d9c36c9b788f6fcb1026559c1bd08f83ca0f55ffb1f3b473955d0fda7742ebbf" exitCode=2 Mar 17 09:13:57 crc kubenswrapper[4813]: I0317 09:13:57.872832 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dpghh" event={"ID":"845ee13b-7ed9-4910-958b-5269812237f0","Type":"ContainerDied","Data":"d9c36c9b788f6fcb1026559c1bd08f83ca0f55ffb1f3b473955d0fda7742ebbf"} Mar 17 09:13:57 crc kubenswrapper[4813]: I0317 09:13:57.873273 4813 scope.go:117] "RemoveContainer" containerID="d9c36c9b788f6fcb1026559c1bd08f83ca0f55ffb1f3b473955d0fda7742ebbf" Mar 17 09:13:59 crc kubenswrapper[4813]: I0317 09:13:59.262074 4813 ???:1] "http: TLS handshake error from 192.168.126.11:46840: no serving certificate available for the kubelet" Mar 17 09:13:59 crc kubenswrapper[4813]: I0317 09:13:59.891059 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-cjn8k" Mar 17 09:14:00 crc kubenswrapper[4813]: I0317 09:14:00.128641 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562314-dwd8b"] Mar 17 09:14:00 crc kubenswrapper[4813]: I0317 09:14:00.129403 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562314-dwd8b" Mar 17 09:14:00 crc kubenswrapper[4813]: I0317 09:14:00.132049 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:14:00 crc kubenswrapper[4813]: I0317 09:14:00.132411 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562314-dwd8b"] Mar 17 09:14:00 crc kubenswrapper[4813]: I0317 09:14:00.273251 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92g4b\" (UniqueName: \"kubernetes.io/projected/9225f4aa-611e-447c-94b7-804e8c973203-kube-api-access-92g4b\") pod \"auto-csr-approver-29562314-dwd8b\" (UID: \"9225f4aa-611e-447c-94b7-804e8c973203\") " pod="openshift-infra/auto-csr-approver-29562314-dwd8b" Mar 17 09:14:00 crc kubenswrapper[4813]: I0317 09:14:00.374404 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92g4b\" (UniqueName: \"kubernetes.io/projected/9225f4aa-611e-447c-94b7-804e8c973203-kube-api-access-92g4b\") pod \"auto-csr-approver-29562314-dwd8b\" (UID: \"9225f4aa-611e-447c-94b7-804e8c973203\") " pod="openshift-infra/auto-csr-approver-29562314-dwd8b" Mar 17 09:14:00 crc kubenswrapper[4813]: I0317 09:14:00.409514 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92g4b\" (UniqueName: \"kubernetes.io/projected/9225f4aa-611e-447c-94b7-804e8c973203-kube-api-access-92g4b\") pod \"auto-csr-approver-29562314-dwd8b\" (UID: \"9225f4aa-611e-447c-94b7-804e8c973203\") " pod="openshift-infra/auto-csr-approver-29562314-dwd8b" Mar 17 09:14:00 crc kubenswrapper[4813]: I0317 09:14:00.448314 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562314-dwd8b" Mar 17 09:14:04 crc kubenswrapper[4813]: I0317 09:14:04.364522 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-xb4k9 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Mar 17 09:14:04 crc kubenswrapper[4813]: I0317 09:14:04.364914 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xb4k9" podUID="3aa0fc32-ed11-482b-9a6d-fcd59544bf8d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Mar 17 09:14:04 crc kubenswrapper[4813]: I0317 09:14:04.364529 4813 patch_prober.go:28] interesting pod/downloads-7954f5f757-xb4k9 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" start-of-body= Mar 17 09:14:04 crc kubenswrapper[4813]: I0317 09:14:04.365335 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-xb4k9" podUID="3aa0fc32-ed11-482b-9a6d-fcd59544bf8d" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.30:8080/\": dial tcp 10.217.0.30:8080: connect: connection refused" Mar 17 09:14:04 crc kubenswrapper[4813]: I0317 09:14:04.422100 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:14:04 crc kubenswrapper[4813]: I0317 09:14:04.427543 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:14:05 crc kubenswrapper[4813]: I0317 09:14:05.932702 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"332581fa-6de2-41c0-9d20-f9d0b269c5df","Type":"ContainerDied","Data":"c9dac6fdab420e1fb6834bbd5d6f4effe83ba541c62f6ab096e6a30c8327930d"} Mar 17 09:14:05 crc kubenswrapper[4813]: I0317 09:14:05.933151 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9dac6fdab420e1fb6834bbd5d6f4effe83ba541c62f6ab096e6a30c8327930d" Mar 17 09:14:05 crc kubenswrapper[4813]: I0317 09:14:05.933864 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6gvsx" event={"ID":"6b3a014a-3574-4c2b-8331-df326d47f240","Type":"ContainerStarted","Data":"e85159594d24618777672cced65c2e1452c4906e3c04bd52403f01727228e6fd"} Mar 17 09:14:05 crc kubenswrapper[4813]: I0317 09:14:05.952315 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 17 09:14:06 crc kubenswrapper[4813]: I0317 09:14:06.067437 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/332581fa-6de2-41c0-9d20-f9d0b269c5df-kubelet-dir\") pod \"332581fa-6de2-41c0-9d20-f9d0b269c5df\" (UID: \"332581fa-6de2-41c0-9d20-f9d0b269c5df\") " Mar 17 09:14:06 crc kubenswrapper[4813]: I0317 09:14:06.067524 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/332581fa-6de2-41c0-9d20-f9d0b269c5df-kube-api-access\") pod \"332581fa-6de2-41c0-9d20-f9d0b269c5df\" (UID: \"332581fa-6de2-41c0-9d20-f9d0b269c5df\") " Mar 17 09:14:06 crc kubenswrapper[4813]: I0317 09:14:06.068805 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/332581fa-6de2-41c0-9d20-f9d0b269c5df-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "332581fa-6de2-41c0-9d20-f9d0b269c5df" (UID: "332581fa-6de2-41c0-9d20-f9d0b269c5df"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:14:06 crc kubenswrapper[4813]: I0317 09:14:06.076211 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/332581fa-6de2-41c0-9d20-f9d0b269c5df-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "332581fa-6de2-41c0-9d20-f9d0b269c5df" (UID: "332581fa-6de2-41c0-9d20-f9d0b269c5df"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:14:06 crc kubenswrapper[4813]: I0317 09:14:06.170020 4813 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/332581fa-6de2-41c0-9d20-f9d0b269c5df-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:06 crc kubenswrapper[4813]: I0317 09:14:06.170087 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/332581fa-6de2-41c0-9d20-f9d0b269c5df-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:06 crc kubenswrapper[4813]: I0317 09:14:06.939094 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 17 09:14:08 crc kubenswrapper[4813]: I0317 09:14:08.554356 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 17 09:14:08 crc kubenswrapper[4813]: I0317 09:14:08.703414 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8f167ed7-d30d-4757-b673-3766ac2a113e-kubelet-dir\") pod \"8f167ed7-d30d-4757-b673-3766ac2a113e\" (UID: \"8f167ed7-d30d-4757-b673-3766ac2a113e\") " Mar 17 09:14:08 crc kubenswrapper[4813]: I0317 09:14:08.703974 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8f167ed7-d30d-4757-b673-3766ac2a113e-kube-api-access\") pod \"8f167ed7-d30d-4757-b673-3766ac2a113e\" (UID: \"8f167ed7-d30d-4757-b673-3766ac2a113e\") " Mar 17 09:14:08 crc kubenswrapper[4813]: I0317 09:14:08.703699 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8f167ed7-d30d-4757-b673-3766ac2a113e-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "8f167ed7-d30d-4757-b673-3766ac2a113e" (UID: "8f167ed7-d30d-4757-b673-3766ac2a113e"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:14:08 crc kubenswrapper[4813]: I0317 09:14:08.712986 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f167ed7-d30d-4757-b673-3766ac2a113e-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "8f167ed7-d30d-4757-b673-3766ac2a113e" (UID: "8f167ed7-d30d-4757-b673-3766ac2a113e"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:14:08 crc kubenswrapper[4813]: I0317 09:14:08.753734 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-bf7m2"] Mar 17 09:14:08 crc kubenswrapper[4813]: I0317 09:14:08.806196 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8f167ed7-d30d-4757-b673-3766ac2a113e-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:08 crc kubenswrapper[4813]: I0317 09:14:08.806246 4813 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/8f167ed7-d30d-4757-b673-3766ac2a113e-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:08 crc kubenswrapper[4813]: I0317 09:14:08.954100 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"8f167ed7-d30d-4757-b673-3766ac2a113e","Type":"ContainerDied","Data":"ed1dbc491bd6599e329285675094273014210d797f3709a3e302215e91b6a874"} Mar 17 09:14:08 crc kubenswrapper[4813]: I0317 09:14:08.954170 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed1dbc491bd6599e329285675094273014210d797f3709a3e302215e91b6a874" Mar 17 09:14:08 crc kubenswrapper[4813]: I0317 09:14:08.954198 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 17 09:14:09 crc kubenswrapper[4813]: I0317 09:14:09.222350 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6ff769775f-cz59s"] Mar 17 09:14:09 crc kubenswrapper[4813]: I0317 09:14:09.222824 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" podUID="1a30c8ff-06c4-478a-bf04-21eb50da1d89" containerName="controller-manager" containerID="cri-o://cbb747b8e38fa24891057d3d949205f36c5a00b26e7e343a7b0d4762b09840fb" gracePeriod=30 Mar 17 09:14:09 crc kubenswrapper[4813]: I0317 09:14:09.237896 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr"] Mar 17 09:14:09 crc kubenswrapper[4813]: I0317 09:14:09.243148 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" podUID="e50bf93b-f84f-4050-b5cf-65733de8db5c" containerName="route-controller-manager" containerID="cri-o://fcd4566bc40b8b2584a7500e2ba1f341d08ae963f3c5c2af9fcb90ca3a290578" gracePeriod=30 Mar 17 09:14:09 crc kubenswrapper[4813]: E0317 09:14:09.834503 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Mar 17 09:14:09 crc kubenswrapper[4813]: E0317 09:14:09.834829 4813 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 17 09:14:09 crc kubenswrapper[4813]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Mar 17 09:14:09 crc kubenswrapper[4813]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-fcghd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29562312-hjpcp_openshift-infra(8626af43-6b5e-49e1-b19f-3986cdecdd2d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Mar 17 09:14:09 crc kubenswrapper[4813]: > logger="UnhandledError" Mar 17 09:14:09 crc kubenswrapper[4813]: E0317 09:14:09.836044 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29562312-hjpcp" podUID="8626af43-6b5e-49e1-b19f-3986cdecdd2d" Mar 17 09:14:09 crc kubenswrapper[4813]: I0317 09:14:09.960335 4813 generic.go:334] "Generic (PLEG): container finished" podID="e50bf93b-f84f-4050-b5cf-65733de8db5c" containerID="fcd4566bc40b8b2584a7500e2ba1f341d08ae963f3c5c2af9fcb90ca3a290578" exitCode=0 Mar 17 09:14:09 crc kubenswrapper[4813]: I0317 09:14:09.960393 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" event={"ID":"e50bf93b-f84f-4050-b5cf-65733de8db5c","Type":"ContainerDied","Data":"fcd4566bc40b8b2584a7500e2ba1f341d08ae963f3c5c2af9fcb90ca3a290578"} Mar 17 09:14:09 crc kubenswrapper[4813]: I0317 09:14:09.963115 4813 generic.go:334] "Generic (PLEG): container finished" podID="1a30c8ff-06c4-478a-bf04-21eb50da1d89" containerID="cbb747b8e38fa24891057d3d949205f36c5a00b26e7e343a7b0d4762b09840fb" exitCode=0 Mar 17 09:14:09 crc kubenswrapper[4813]: I0317 09:14:09.963195 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" event={"ID":"1a30c8ff-06c4-478a-bf04-21eb50da1d89","Type":"ContainerDied","Data":"cbb747b8e38fa24891057d3d949205f36c5a00b26e7e343a7b0d4762b09840fb"} Mar 17 09:14:09 crc kubenswrapper[4813]: E0317 09:14:09.966064 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29562312-hjpcp" podUID="8626af43-6b5e-49e1-b19f-3986cdecdd2d" Mar 17 09:14:10 crc kubenswrapper[4813]: W0317 09:14:10.088340 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode49eaa7a_38bf_4996_ad34_64959cbccda2.slice/crio-9fdffdd82d65d337053a38a5ed253a94699ac97963d750f597c483bca38d1043 WatchSource:0}: Error finding container 9fdffdd82d65d337053a38a5ed253a94699ac97963d750f597c483bca38d1043: Status 404 returned error can't find the container with id 9fdffdd82d65d337053a38a5ed253a94699ac97963d750f597c483bca38d1043 Mar 17 09:14:10 crc kubenswrapper[4813]: I0317 09:14:10.294563 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562314-dwd8b"] Mar 17 09:14:10 crc kubenswrapper[4813]: I0317 09:14:10.971003 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bf7m2" event={"ID":"e49eaa7a-38bf-4996-ad34-64959cbccda2","Type":"ContainerStarted","Data":"9fdffdd82d65d337053a38a5ed253a94699ac97963d750f597c483bca38d1043"} Mar 17 09:14:11 crc kubenswrapper[4813]: I0317 09:14:11.327471 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:14:11 crc kubenswrapper[4813]: I0317 09:14:11.993514 4813 patch_prober.go:28] interesting pod/controller-manager-6ff769775f-cz59s container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.45:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 09:14:11 crc kubenswrapper[4813]: I0317 09:14:11.993864 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" podUID="1a30c8ff-06c4-478a-bf04-21eb50da1d89" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.45:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 09:14:12 crc kubenswrapper[4813]: I0317 09:14:12.009482 4813 patch_prober.go:28] interesting pod/route-controller-manager-5885dd695f-qh6zr container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.46:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 17 09:14:12 crc kubenswrapper[4813]: I0317 09:14:12.009572 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" podUID="e50bf93b-f84f-4050-b5cf-65733de8db5c" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.46:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 17 09:14:13 crc kubenswrapper[4813]: W0317 09:14:13.458877 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9225f4aa_611e_447c_94b7_804e8c973203.slice/crio-70c2f8afa65a43ca9fe1af63181535c2a651816a5f44046db560f4716955f0cf WatchSource:0}: Error finding container 70c2f8afa65a43ca9fe1af63181535c2a651816a5f44046db560f4716955f0cf: Status 404 returned error can't find the container with id 70c2f8afa65a43ca9fe1af63181535c2a651816a5f44046db560f4716955f0cf Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.501805 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.536377 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-74f9cc6d4d-894cd"] Mar 17 09:14:13 crc kubenswrapper[4813]: E0317 09:14:13.536997 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a30c8ff-06c4-478a-bf04-21eb50da1d89" containerName="controller-manager" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.537020 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a30c8ff-06c4-478a-bf04-21eb50da1d89" containerName="controller-manager" Mar 17 09:14:13 crc kubenswrapper[4813]: E0317 09:14:13.537036 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f167ed7-d30d-4757-b673-3766ac2a113e" containerName="pruner" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.537046 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f167ed7-d30d-4757-b673-3766ac2a113e" containerName="pruner" Mar 17 09:14:13 crc kubenswrapper[4813]: E0317 09:14:13.537054 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="332581fa-6de2-41c0-9d20-f9d0b269c5df" containerName="pruner" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.537061 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="332581fa-6de2-41c0-9d20-f9d0b269c5df" containerName="pruner" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.537191 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="332581fa-6de2-41c0-9d20-f9d0b269c5df" containerName="pruner" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.537207 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a30c8ff-06c4-478a-bf04-21eb50da1d89" containerName="controller-manager" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.537222 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f167ed7-d30d-4757-b673-3766ac2a113e" containerName="pruner" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.537820 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.541560 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74f9cc6d4d-894cd"] Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.548974 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.692732 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e50bf93b-f84f-4050-b5cf-65733de8db5c-config\") pod \"e50bf93b-f84f-4050-b5cf-65733de8db5c\" (UID: \"e50bf93b-f84f-4050-b5cf-65733de8db5c\") " Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.692776 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e50bf93b-f84f-4050-b5cf-65733de8db5c-serving-cert\") pod \"e50bf93b-f84f-4050-b5cf-65733de8db5c\" (UID: \"e50bf93b-f84f-4050-b5cf-65733de8db5c\") " Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.692859 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdlsg\" (UniqueName: \"kubernetes.io/projected/e50bf93b-f84f-4050-b5cf-65733de8db5c-kube-api-access-jdlsg\") pod \"e50bf93b-f84f-4050-b5cf-65733de8db5c\" (UID: \"e50bf93b-f84f-4050-b5cf-65733de8db5c\") " Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.692884 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a30c8ff-06c4-478a-bf04-21eb50da1d89-client-ca\") pod \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\" (UID: \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\") " Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.692922 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e50bf93b-f84f-4050-b5cf-65733de8db5c-client-ca\") pod \"e50bf93b-f84f-4050-b5cf-65733de8db5c\" (UID: \"e50bf93b-f84f-4050-b5cf-65733de8db5c\") " Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.692950 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blwtx\" (UniqueName: \"kubernetes.io/projected/1a30c8ff-06c4-478a-bf04-21eb50da1d89-kube-api-access-blwtx\") pod \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\" (UID: \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\") " Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.692969 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a30c8ff-06c4-478a-bf04-21eb50da1d89-serving-cert\") pod \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\" (UID: \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\") " Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.692999 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a30c8ff-06c4-478a-bf04-21eb50da1d89-config\") pod \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\" (UID: \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\") " Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.693032 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1a30c8ff-06c4-478a-bf04-21eb50da1d89-proxy-ca-bundles\") pod \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\" (UID: \"1a30c8ff-06c4-478a-bf04-21eb50da1d89\") " Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.693302 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-client-ca\") pod \"controller-manager-74f9cc6d4d-894cd\" (UID: \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\") " pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.693352 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfnzm\" (UniqueName: \"kubernetes.io/projected/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-kube-api-access-xfnzm\") pod \"controller-manager-74f9cc6d4d-894cd\" (UID: \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\") " pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.693378 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-config\") pod \"controller-manager-74f9cc6d4d-894cd\" (UID: \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\") " pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.693428 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-proxy-ca-bundles\") pod \"controller-manager-74f9cc6d4d-894cd\" (UID: \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\") " pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.693451 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-serving-cert\") pod \"controller-manager-74f9cc6d4d-894cd\" (UID: \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\") " pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.693881 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a30c8ff-06c4-478a-bf04-21eb50da1d89-client-ca" (OuterVolumeSpecName: "client-ca") pod "1a30c8ff-06c4-478a-bf04-21eb50da1d89" (UID: "1a30c8ff-06c4-478a-bf04-21eb50da1d89"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.693909 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a30c8ff-06c4-478a-bf04-21eb50da1d89-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1a30c8ff-06c4-478a-bf04-21eb50da1d89" (UID: "1a30c8ff-06c4-478a-bf04-21eb50da1d89"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.693956 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a30c8ff-06c4-478a-bf04-21eb50da1d89-config" (OuterVolumeSpecName: "config") pod "1a30c8ff-06c4-478a-bf04-21eb50da1d89" (UID: "1a30c8ff-06c4-478a-bf04-21eb50da1d89"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.695412 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e50bf93b-f84f-4050-b5cf-65733de8db5c-client-ca" (OuterVolumeSpecName: "client-ca") pod "e50bf93b-f84f-4050-b5cf-65733de8db5c" (UID: "e50bf93b-f84f-4050-b5cf-65733de8db5c"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.695474 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e50bf93b-f84f-4050-b5cf-65733de8db5c-config" (OuterVolumeSpecName: "config") pod "e50bf93b-f84f-4050-b5cf-65733de8db5c" (UID: "e50bf93b-f84f-4050-b5cf-65733de8db5c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.698892 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a30c8ff-06c4-478a-bf04-21eb50da1d89-kube-api-access-blwtx" (OuterVolumeSpecName: "kube-api-access-blwtx") pod "1a30c8ff-06c4-478a-bf04-21eb50da1d89" (UID: "1a30c8ff-06c4-478a-bf04-21eb50da1d89"). InnerVolumeSpecName "kube-api-access-blwtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.699849 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e50bf93b-f84f-4050-b5cf-65733de8db5c-kube-api-access-jdlsg" (OuterVolumeSpecName: "kube-api-access-jdlsg") pod "e50bf93b-f84f-4050-b5cf-65733de8db5c" (UID: "e50bf93b-f84f-4050-b5cf-65733de8db5c"). InnerVolumeSpecName "kube-api-access-jdlsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.700457 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e50bf93b-f84f-4050-b5cf-65733de8db5c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e50bf93b-f84f-4050-b5cf-65733de8db5c" (UID: "e50bf93b-f84f-4050-b5cf-65733de8db5c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.701518 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1a30c8ff-06c4-478a-bf04-21eb50da1d89-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1a30c8ff-06c4-478a-bf04-21eb50da1d89" (UID: "1a30c8ff-06c4-478a-bf04-21eb50da1d89"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.796150 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-proxy-ca-bundles\") pod \"controller-manager-74f9cc6d4d-894cd\" (UID: \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\") " pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.796199 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-serving-cert\") pod \"controller-manager-74f9cc6d4d-894cd\" (UID: \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\") " pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.796250 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-client-ca\") pod \"controller-manager-74f9cc6d4d-894cd\" (UID: \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\") " pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.796286 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfnzm\" (UniqueName: \"kubernetes.io/projected/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-kube-api-access-xfnzm\") pod \"controller-manager-74f9cc6d4d-894cd\" (UID: \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\") " pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.796303 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-config\") pod \"controller-manager-74f9cc6d4d-894cd\" (UID: \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\") " pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.796395 4813 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1a30c8ff-06c4-478a-bf04-21eb50da1d89-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.796409 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e50bf93b-f84f-4050-b5cf-65733de8db5c-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.796420 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e50bf93b-f84f-4050-b5cf-65733de8db5c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.796432 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdlsg\" (UniqueName: \"kubernetes.io/projected/e50bf93b-f84f-4050-b5cf-65733de8db5c-kube-api-access-jdlsg\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.796445 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1a30c8ff-06c4-478a-bf04-21eb50da1d89-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.796457 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e50bf93b-f84f-4050-b5cf-65733de8db5c-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.796467 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blwtx\" (UniqueName: \"kubernetes.io/projected/1a30c8ff-06c4-478a-bf04-21eb50da1d89-kube-api-access-blwtx\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.796475 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1a30c8ff-06c4-478a-bf04-21eb50da1d89-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.796512 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1a30c8ff-06c4-478a-bf04-21eb50da1d89-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.799126 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-client-ca\") pod \"controller-manager-74f9cc6d4d-894cd\" (UID: \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\") " pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.799146 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-proxy-ca-bundles\") pod \"controller-manager-74f9cc6d4d-894cd\" (UID: \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\") " pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.799197 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-config\") pod \"controller-manager-74f9cc6d4d-894cd\" (UID: \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\") " pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.803045 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-serving-cert\") pod \"controller-manager-74f9cc6d4d-894cd\" (UID: \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\") " pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.816893 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfnzm\" (UniqueName: \"kubernetes.io/projected/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-kube-api-access-xfnzm\") pod \"controller-manager-74f9cc6d4d-894cd\" (UID: \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\") " pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.869583 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.987980 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" event={"ID":"1a30c8ff-06c4-478a-bf04-21eb50da1d89","Type":"ContainerDied","Data":"bee449c8ac9f58cd0f1cc43df8051ece39ec25f0129d0616913fc19ee6e6f89f"} Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.988037 4813 scope.go:117] "RemoveContainer" containerID="cbb747b8e38fa24891057d3d949205f36c5a00b26e7e343a7b0d4762b09840fb" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.988141 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6ff769775f-cz59s" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.994579 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562314-dwd8b" event={"ID":"9225f4aa-611e-447c-94b7-804e8c973203","Type":"ContainerStarted","Data":"70c2f8afa65a43ca9fe1af63181535c2a651816a5f44046db560f4716955f0cf"} Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.997271 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-cluster-samples-operator_cluster-samples-operator-665b6dd947-dpghh_845ee13b-7ed9-4910-958b-5269812237f0/cluster-samples-operator/0.log" Mar 17 09:14:13 crc kubenswrapper[4813]: I0317 09:14:13.997357 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-dpghh" event={"ID":"845ee13b-7ed9-4910-958b-5269812237f0","Type":"ContainerStarted","Data":"45674d02a46a7979fe5e028b1ee74ec4425b4b9fb65bf256a0a9a48e58bfdff0"} Mar 17 09:14:14 crc kubenswrapper[4813]: I0317 09:14:13.999619 4813 generic.go:334] "Generic (PLEG): container finished" podID="6b3a014a-3574-4c2b-8331-df326d47f240" containerID="d6d775b3a82cedca3c82e251809e2d700188f2a218b26d154e599e0ab0607c24" exitCode=0 Mar 17 09:14:14 crc kubenswrapper[4813]: I0317 09:14:13.999712 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6gvsx" event={"ID":"6b3a014a-3574-4c2b-8331-df326d47f240","Type":"ContainerDied","Data":"d6d775b3a82cedca3c82e251809e2d700188f2a218b26d154e599e0ab0607c24"} Mar 17 09:14:14 crc kubenswrapper[4813]: I0317 09:14:14.002548 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" event={"ID":"e50bf93b-f84f-4050-b5cf-65733de8db5c","Type":"ContainerDied","Data":"aa730b8da6c17173b497c1f2cd8704f3fe4819c17a6bf449ecd2e4f102aea7f2"} Mar 17 09:14:14 crc kubenswrapper[4813]: I0317 09:14:14.002683 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr" Mar 17 09:14:14 crc kubenswrapper[4813]: I0317 09:14:14.062679 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-6ff769775f-cz59s"] Mar 17 09:14:14 crc kubenswrapper[4813]: I0317 09:14:14.069458 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-6ff769775f-cz59s"] Mar 17 09:14:14 crc kubenswrapper[4813]: I0317 09:14:14.072154 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr"] Mar 17 09:14:14 crc kubenswrapper[4813]: I0317 09:14:14.074515 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5885dd695f-qh6zr"] Mar 17 09:14:14 crc kubenswrapper[4813]: I0317 09:14:14.113705 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:14:14 crc kubenswrapper[4813]: I0317 09:14:14.113755 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:14:14 crc kubenswrapper[4813]: I0317 09:14:14.381817 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-xb4k9" Mar 17 09:14:14 crc kubenswrapper[4813]: I0317 09:14:14.736547 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a30c8ff-06c4-478a-bf04-21eb50da1d89" path="/var/lib/kubelet/pods/1a30c8ff-06c4-478a-bf04-21eb50da1d89/volumes" Mar 17 09:14:14 crc kubenswrapper[4813]: I0317 09:14:14.737363 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e50bf93b-f84f-4050-b5cf-65733de8db5c" path="/var/lib/kubelet/pods/e50bf93b-f84f-4050-b5cf-65733de8db5c/volumes" Mar 17 09:14:15 crc kubenswrapper[4813]: I0317 09:14:15.929642 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb"] Mar 17 09:14:15 crc kubenswrapper[4813]: E0317 09:14:15.930338 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e50bf93b-f84f-4050-b5cf-65733de8db5c" containerName="route-controller-manager" Mar 17 09:14:15 crc kubenswrapper[4813]: I0317 09:14:15.930350 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e50bf93b-f84f-4050-b5cf-65733de8db5c" containerName="route-controller-manager" Mar 17 09:14:15 crc kubenswrapper[4813]: I0317 09:14:15.931895 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e50bf93b-f84f-4050-b5cf-65733de8db5c" containerName="route-controller-manager" Mar 17 09:14:15 crc kubenswrapper[4813]: I0317 09:14:15.932340 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" Mar 17 09:14:15 crc kubenswrapper[4813]: I0317 09:14:15.933403 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb"] Mar 17 09:14:15 crc kubenswrapper[4813]: I0317 09:14:15.934685 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 17 09:14:15 crc kubenswrapper[4813]: I0317 09:14:15.935145 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 17 09:14:15 crc kubenswrapper[4813]: I0317 09:14:15.935361 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 17 09:14:15 crc kubenswrapper[4813]: I0317 09:14:15.935476 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 17 09:14:15 crc kubenswrapper[4813]: I0317 09:14:15.935508 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 17 09:14:15 crc kubenswrapper[4813]: I0317 09:14:15.935796 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 17 09:14:16 crc kubenswrapper[4813]: I0317 09:14:16.125585 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-client-ca\") pod \"route-controller-manager-647b8ff8cf-w5vvb\" (UID: \"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44\") " pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" Mar 17 09:14:16 crc kubenswrapper[4813]: I0317 09:14:16.125878 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7kdd\" (UniqueName: \"kubernetes.io/projected/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-kube-api-access-m7kdd\") pod \"route-controller-manager-647b8ff8cf-w5vvb\" (UID: \"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44\") " pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" Mar 17 09:14:16 crc kubenswrapper[4813]: I0317 09:14:16.126074 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-config\") pod \"route-controller-manager-647b8ff8cf-w5vvb\" (UID: \"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44\") " pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" Mar 17 09:14:16 crc kubenswrapper[4813]: I0317 09:14:16.126213 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-serving-cert\") pod \"route-controller-manager-647b8ff8cf-w5vvb\" (UID: \"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44\") " pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" Mar 17 09:14:16 crc kubenswrapper[4813]: I0317 09:14:16.226937 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-config\") pod \"route-controller-manager-647b8ff8cf-w5vvb\" (UID: \"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44\") " pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" Mar 17 09:14:16 crc kubenswrapper[4813]: I0317 09:14:16.227003 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-serving-cert\") pod \"route-controller-manager-647b8ff8cf-w5vvb\" (UID: \"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44\") " pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" Mar 17 09:14:16 crc kubenswrapper[4813]: I0317 09:14:16.227030 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-client-ca\") pod \"route-controller-manager-647b8ff8cf-w5vvb\" (UID: \"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44\") " pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" Mar 17 09:14:16 crc kubenswrapper[4813]: I0317 09:14:16.227057 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7kdd\" (UniqueName: \"kubernetes.io/projected/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-kube-api-access-m7kdd\") pod \"route-controller-manager-647b8ff8cf-w5vvb\" (UID: \"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44\") " pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" Mar 17 09:14:16 crc kubenswrapper[4813]: I0317 09:14:16.228350 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-client-ca\") pod \"route-controller-manager-647b8ff8cf-w5vvb\" (UID: \"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44\") " pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" Mar 17 09:14:16 crc kubenswrapper[4813]: I0317 09:14:16.230566 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-config\") pod \"route-controller-manager-647b8ff8cf-w5vvb\" (UID: \"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44\") " pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" Mar 17 09:14:16 crc kubenswrapper[4813]: I0317 09:14:16.244899 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-serving-cert\") pod \"route-controller-manager-647b8ff8cf-w5vvb\" (UID: \"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44\") " pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" Mar 17 09:14:16 crc kubenswrapper[4813]: I0317 09:14:16.248670 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7kdd\" (UniqueName: \"kubernetes.io/projected/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-kube-api-access-m7kdd\") pod \"route-controller-manager-647b8ff8cf-w5vvb\" (UID: \"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44\") " pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" Mar 17 09:14:16 crc kubenswrapper[4813]: I0317 09:14:16.253410 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" Mar 17 09:14:18 crc kubenswrapper[4813]: I0317 09:14:18.923703 4813 scope.go:117] "RemoveContainer" containerID="fcd4566bc40b8b2584a7500e2ba1f341d08ae963f3c5c2af9fcb90ca3a290578" Mar 17 09:14:18 crc kubenswrapper[4813]: E0317 09:14:18.949118 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 17 09:14:18 crc kubenswrapper[4813]: E0317 09:14:18.949297 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-q7vwj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-5lhdz_openshift-marketplace(404fc67e-83be-4be6-bf6e-4279c0543316): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 17 09:14:18 crc kubenswrapper[4813]: E0317 09:14:18.950626 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-5lhdz" podUID="404fc67e-83be-4be6-bf6e-4279c0543316" Mar 17 09:14:18 crc kubenswrapper[4813]: E0317 09:14:18.978479 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 17 09:14:18 crc kubenswrapper[4813]: E0317 09:14:18.978673 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tt5rs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-lxrc7_openshift-marketplace(1b9e1de0-da0f-422e-a79b-4fb54d2e3acb): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 17 09:14:18 crc kubenswrapper[4813]: E0317 09:14:18.979932 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-lxrc7" podUID="1b9e1de0-da0f-422e-a79b-4fb54d2e3acb" Mar 17 09:14:19 crc kubenswrapper[4813]: E0317 09:14:19.042087 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-5lhdz" podUID="404fc67e-83be-4be6-bf6e-4279c0543316" Mar 17 09:14:19 crc kubenswrapper[4813]: E0317 09:14:19.042589 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-lxrc7" podUID="1b9e1de0-da0f-422e-a79b-4fb54d2e3acb" Mar 17 09:14:19 crc kubenswrapper[4813]: I0317 09:14:19.420018 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb"] Mar 17 09:14:19 crc kubenswrapper[4813]: W0317 09:14:19.423353 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc9de28e_cda4_4f04_a6c1_64c7ee2d2e44.slice/crio-cb570d829effa0a42fb5c2ed04213ea60556e7a604a01fa205c5ea2b73fa7ed5 WatchSource:0}: Error finding container cb570d829effa0a42fb5c2ed04213ea60556e7a604a01fa205c5ea2b73fa7ed5: Status 404 returned error can't find the container with id cb570d829effa0a42fb5c2ed04213ea60556e7a604a01fa205c5ea2b73fa7ed5 Mar 17 09:14:19 crc kubenswrapper[4813]: I0317 09:14:19.467785 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74f9cc6d4d-894cd"] Mar 17 09:14:19 crc kubenswrapper[4813]: W0317 09:14:19.485186 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7d000133_b0b4_43c8_b7b5_6f3b6de1a3f0.slice/crio-7152088bb0ccf53b1a9d72144cecc214a6703ddf69872e47e8903c8ce412393c WatchSource:0}: Error finding container 7152088bb0ccf53b1a9d72144cecc214a6703ddf69872e47e8903c8ce412393c: Status 404 returned error can't find the container with id 7152088bb0ccf53b1a9d72144cecc214a6703ddf69872e47e8903c8ce412393c Mar 17 09:14:19 crc kubenswrapper[4813]: I0317 09:14:19.774695 4813 ???:1] "http: TLS handshake error from 192.168.126.11:44316: no serving certificate available for the kubelet" Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.038676 4813 generic.go:334] "Generic (PLEG): container finished" podID="12cf87e0-5d1a-4598-bd14-e3ef92a56dc1" containerID="1f30c4c96c804a7b9dc9a6b5d7a7558fc2822e74e875b763607ddb261b75d90b" exitCode=0 Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.038794 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-js422" event={"ID":"12cf87e0-5d1a-4598-bd14-e3ef92a56dc1","Type":"ContainerDied","Data":"1f30c4c96c804a7b9dc9a6b5d7a7558fc2822e74e875b763607ddb261b75d90b"} Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.043188 4813 generic.go:334] "Generic (PLEG): container finished" podID="e49eaa7a-38bf-4996-ad34-64959cbccda2" containerID="27d9c820b252b46f72190dc09d97f814a6827d85e54fe13cb7ebc5a56f9b27f0" exitCode=0 Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.043325 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bf7m2" event={"ID":"e49eaa7a-38bf-4996-ad34-64959cbccda2","Type":"ContainerDied","Data":"27d9c820b252b46f72190dc09d97f814a6827d85e54fe13cb7ebc5a56f9b27f0"} Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.052303 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" event={"ID":"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0","Type":"ContainerStarted","Data":"f226df3be2e759f1108da445677a823bda3cd97b637d30f9184454fbf25d23f6"} Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.052353 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" event={"ID":"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0","Type":"ContainerStarted","Data":"7152088bb0ccf53b1a9d72144cecc214a6703ddf69872e47e8903c8ce412393c"} Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.052739 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.054537 4813 generic.go:334] "Generic (PLEG): container finished" podID="65af6b35-41bd-4ad2-b8d4-0b653cb18dcf" containerID="f10efea5d3325a7eeeeb9a544f307de5d03087d18da00fc581580837e1f39274" exitCode=0 Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.054616 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2skjk" event={"ID":"65af6b35-41bd-4ad2-b8d4-0b653cb18dcf","Type":"ContainerDied","Data":"f10efea5d3325a7eeeeb9a544f307de5d03087d18da00fc581580837e1f39274"} Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.059091 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.068692 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5dpjf" event={"ID":"b7d1b948-0f9a-4ba3-9f58-88590e13ba63","Type":"ContainerDied","Data":"2feaf20e554846b8e9aa047cc5fe61f1c3a7b837e0fc677962aee650e02877b8"} Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.068750 4813 generic.go:334] "Generic (PLEG): container finished" podID="b7d1b948-0f9a-4ba3-9f58-88590e13ba63" containerID="2feaf20e554846b8e9aa047cc5fe61f1c3a7b837e0fc677962aee650e02877b8" exitCode=0 Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.082450 4813 generic.go:334] "Generic (PLEG): container finished" podID="4ca1cd05-e8fc-47b7-904b-a0a544f654b0" containerID="ba6c548eb755d9564ab731194956b862e51e6e871050ad96f7f70efec7ed5611" exitCode=0 Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.082536 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x95rg" event={"ID":"4ca1cd05-e8fc-47b7-904b-a0a544f654b0","Type":"ContainerDied","Data":"ba6c548eb755d9564ab731194956b862e51e6e871050ad96f7f70efec7ed5611"} Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.084736 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562314-dwd8b" event={"ID":"9225f4aa-611e-447c-94b7-804e8c973203","Type":"ContainerStarted","Data":"eb525fa6ddd5e7d38d337ed03384d69605ce16132833739caba5ba0ad5e4edab"} Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.088456 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" podStartSLOduration=11.08844042 podStartE2EDuration="11.08844042s" podCreationTimestamp="2026-03-17 09:14:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:14:20.08785105 +0000 UTC m=+282.188654549" watchObservedRunningTime="2026-03-17 09:14:20.08844042 +0000 UTC m=+282.189243919" Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.100063 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" event={"ID":"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44","Type":"ContainerStarted","Data":"a1ffec25ed32dda25b651929fec928c8bb12892cc5e8e71de10dc031edd0a35e"} Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.100103 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" event={"ID":"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44","Type":"ContainerStarted","Data":"cb570d829effa0a42fb5c2ed04213ea60556e7a604a01fa205c5ea2b73fa7ed5"} Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.100406 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.165012 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562314-dwd8b" podStartSLOduration=14.564024697 podStartE2EDuration="20.164992588s" podCreationTimestamp="2026-03-17 09:14:00 +0000 UTC" firstStartedPulling="2026-03-17 09:14:13.461662115 +0000 UTC m=+275.562465614" lastFinishedPulling="2026-03-17 09:14:19.062630006 +0000 UTC m=+281.163433505" observedRunningTime="2026-03-17 09:14:20.164961606 +0000 UTC m=+282.265765105" watchObservedRunningTime="2026-03-17 09:14:20.164992588 +0000 UTC m=+282.265796097" Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.191860 4813 csr.go:261] certificate signing request csr-dsltp is approved, waiting to be issued Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.195687 4813 csr.go:257] certificate signing request csr-dsltp is issued Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.205019 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" podStartSLOduration=11.205003449 podStartE2EDuration="11.205003449s" podCreationTimestamp="2026-03-17 09:14:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:14:20.20176484 +0000 UTC m=+282.302568339" watchObservedRunningTime="2026-03-17 09:14:20.205003449 +0000 UTC m=+282.305806948" Mar 17 09:14:20 crc kubenswrapper[4813]: I0317 09:14:20.234746 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" Mar 17 09:14:21 crc kubenswrapper[4813]: I0317 09:14:21.111562 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5dpjf" event={"ID":"b7d1b948-0f9a-4ba3-9f58-88590e13ba63","Type":"ContainerStarted","Data":"97ebdfe561e724beec2af09e7bff0a01b83678538da9176586cefd595e83d7ca"} Mar 17 09:14:21 crc kubenswrapper[4813]: I0317 09:14:21.116051 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-js422" event={"ID":"12cf87e0-5d1a-4598-bd14-e3ef92a56dc1","Type":"ContainerStarted","Data":"4e2f9e427afe63058f581dbffd89d35960f8c1034239db65888db4a7b6558688"} Mar 17 09:14:21 crc kubenswrapper[4813]: I0317 09:14:21.122136 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x95rg" event={"ID":"4ca1cd05-e8fc-47b7-904b-a0a544f654b0","Type":"ContainerStarted","Data":"988ee7f8b66efd6561f1e628375cfe7016af93e9579b928a9f95754744bb9ac7"} Mar 17 09:14:21 crc kubenswrapper[4813]: I0317 09:14:21.125388 4813 generic.go:334] "Generic (PLEG): container finished" podID="9225f4aa-611e-447c-94b7-804e8c973203" containerID="eb525fa6ddd5e7d38d337ed03384d69605ce16132833739caba5ba0ad5e4edab" exitCode=0 Mar 17 09:14:21 crc kubenswrapper[4813]: I0317 09:14:21.125458 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562314-dwd8b" event={"ID":"9225f4aa-611e-447c-94b7-804e8c973203","Type":"ContainerDied","Data":"eb525fa6ddd5e7d38d337ed03384d69605ce16132833739caba5ba0ad5e4edab"} Mar 17 09:14:21 crc kubenswrapper[4813]: I0317 09:14:21.128766 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2skjk" event={"ID":"65af6b35-41bd-4ad2-b8d4-0b653cb18dcf","Type":"ContainerStarted","Data":"e61a91f1f051970c805f6541cb76afeeb6855918e43560a4f856e2e03dcf12d9"} Mar 17 09:14:21 crc kubenswrapper[4813]: I0317 09:14:21.128851 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5dpjf" podStartSLOduration=3.105288873 podStartE2EDuration="30.128836531s" podCreationTimestamp="2026-03-17 09:13:51 +0000 UTC" firstStartedPulling="2026-03-17 09:13:53.639980957 +0000 UTC m=+255.740784456" lastFinishedPulling="2026-03-17 09:14:20.663528615 +0000 UTC m=+282.764332114" observedRunningTime="2026-03-17 09:14:21.126111958 +0000 UTC m=+283.226915457" watchObservedRunningTime="2026-03-17 09:14:21.128836531 +0000 UTC m=+283.229640040" Mar 17 09:14:21 crc kubenswrapper[4813]: I0317 09:14:21.152854 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-js422" podStartSLOduration=2.012901942 podStartE2EDuration="28.152838859s" podCreationTimestamp="2026-03-17 09:13:53 +0000 UTC" firstStartedPulling="2026-03-17 09:13:54.683812134 +0000 UTC m=+256.784615633" lastFinishedPulling="2026-03-17 09:14:20.823749051 +0000 UTC m=+282.924552550" observedRunningTime="2026-03-17 09:14:21.150864482 +0000 UTC m=+283.251667991" watchObservedRunningTime="2026-03-17 09:14:21.152838859 +0000 UTC m=+283.253642358" Mar 17 09:14:21 crc kubenswrapper[4813]: I0317 09:14:21.186305 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-x95rg" podStartSLOduration=12.49525287 podStartE2EDuration="28.186289258s" podCreationTimestamp="2026-03-17 09:13:53 +0000 UTC" firstStartedPulling="2026-03-17 09:14:04.926838562 +0000 UTC m=+267.027642071" lastFinishedPulling="2026-03-17 09:14:20.61787496 +0000 UTC m=+282.718678459" observedRunningTime="2026-03-17 09:14:21.18520846 +0000 UTC m=+283.286011959" watchObservedRunningTime="2026-03-17 09:14:21.186289258 +0000 UTC m=+283.287092757" Mar 17 09:14:21 crc kubenswrapper[4813]: I0317 09:14:21.196892 4813 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-24 15:40:08.65368229 +0000 UTC Mar 17 09:14:21 crc kubenswrapper[4813]: I0317 09:14:21.197378 4813 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6774h25m47.456307775s for next certificate rotation Mar 17 09:14:21 crc kubenswrapper[4813]: I0317 09:14:21.209324 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2skjk" podStartSLOduration=2.046703034 podStartE2EDuration="30.209289001s" podCreationTimestamp="2026-03-17 09:13:51 +0000 UTC" firstStartedPulling="2026-03-17 09:13:52.563482087 +0000 UTC m=+254.664285586" lastFinishedPulling="2026-03-17 09:14:20.726068044 +0000 UTC m=+282.826871553" observedRunningTime="2026-03-17 09:14:21.205857424 +0000 UTC m=+283.306660933" watchObservedRunningTime="2026-03-17 09:14:21.209289001 +0000 UTC m=+283.310092500" Mar 17 09:14:21 crc kubenswrapper[4813]: I0317 09:14:21.451870 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2skjk" Mar 17 09:14:21 crc kubenswrapper[4813]: I0317 09:14:21.451921 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2skjk" Mar 17 09:14:21 crc kubenswrapper[4813]: I0317 09:14:21.873208 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5dpjf" Mar 17 09:14:21 crc kubenswrapper[4813]: I0317 09:14:21.873250 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5dpjf" Mar 17 09:14:22 crc kubenswrapper[4813]: I0317 09:14:22.198055 4813 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-22 23:13:52.987298648 +0000 UTC Mar 17 09:14:22 crc kubenswrapper[4813]: I0317 09:14:22.198097 4813 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6733h59m30.789205264s for next certificate rotation Mar 17 09:14:22 crc kubenswrapper[4813]: I0317 09:14:22.676352 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-2skjk" podUID="65af6b35-41bd-4ad2-b8d4-0b653cb18dcf" containerName="registry-server" probeResult="failure" output=< Mar 17 09:14:22 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Mar 17 09:14:22 crc kubenswrapper[4813]: > Mar 17 09:14:22 crc kubenswrapper[4813]: I0317 09:14:22.920922 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-5dpjf" podUID="b7d1b948-0f9a-4ba3-9f58-88590e13ba63" containerName="registry-server" probeResult="failure" output=< Mar 17 09:14:22 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Mar 17 09:14:22 crc kubenswrapper[4813]: > Mar 17 09:14:23 crc kubenswrapper[4813]: I0317 09:14:23.485172 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-js422" Mar 17 09:14:23 crc kubenswrapper[4813]: I0317 09:14:23.485214 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-js422" Mar 17 09:14:23 crc kubenswrapper[4813]: I0317 09:14:23.531210 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-js422" Mar 17 09:14:23 crc kubenswrapper[4813]: I0317 09:14:23.886454 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-x95rg" Mar 17 09:14:23 crc kubenswrapper[4813]: I0317 09:14:23.886765 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-x95rg" Mar 17 09:14:23 crc kubenswrapper[4813]: I0317 09:14:23.993631 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-x95rg" Mar 17 09:14:24 crc kubenswrapper[4813]: I0317 09:14:24.571089 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-7w9l5" Mar 17 09:14:25 crc kubenswrapper[4813]: I0317 09:14:25.282799 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 17 09:14:25 crc kubenswrapper[4813]: I0317 09:14:25.283458 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 17 09:14:25 crc kubenswrapper[4813]: I0317 09:14:25.285141 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 17 09:14:25 crc kubenswrapper[4813]: I0317 09:14:25.285358 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 17 09:14:25 crc kubenswrapper[4813]: I0317 09:14:25.286400 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 17 09:14:25 crc kubenswrapper[4813]: I0317 09:14:25.456446 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 17 09:14:25 crc kubenswrapper[4813]: I0317 09:14:25.456501 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 17 09:14:25 crc kubenswrapper[4813]: I0317 09:14:25.558497 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 17 09:14:25 crc kubenswrapper[4813]: I0317 09:14:25.558571 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 17 09:14:25 crc kubenswrapper[4813]: I0317 09:14:25.558725 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 17 09:14:25 crc kubenswrapper[4813]: I0317 09:14:25.583370 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 17 09:14:25 crc kubenswrapper[4813]: I0317 09:14:25.610903 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 17 09:14:27 crc kubenswrapper[4813]: I0317 09:14:27.920823 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562314-dwd8b" Mar 17 09:14:28 crc kubenswrapper[4813]: I0317 09:14:28.090552 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92g4b\" (UniqueName: \"kubernetes.io/projected/9225f4aa-611e-447c-94b7-804e8c973203-kube-api-access-92g4b\") pod \"9225f4aa-611e-447c-94b7-804e8c973203\" (UID: \"9225f4aa-611e-447c-94b7-804e8c973203\") " Mar 17 09:14:28 crc kubenswrapper[4813]: I0317 09:14:28.095905 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9225f4aa-611e-447c-94b7-804e8c973203-kube-api-access-92g4b" (OuterVolumeSpecName: "kube-api-access-92g4b") pod "9225f4aa-611e-447c-94b7-804e8c973203" (UID: "9225f4aa-611e-447c-94b7-804e8c973203"). InnerVolumeSpecName "kube-api-access-92g4b". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:14:28 crc kubenswrapper[4813]: I0317 09:14:28.180525 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562314-dwd8b" event={"ID":"9225f4aa-611e-447c-94b7-804e8c973203","Type":"ContainerDied","Data":"70c2f8afa65a43ca9fe1af63181535c2a651816a5f44046db560f4716955f0cf"} Mar 17 09:14:28 crc kubenswrapper[4813]: I0317 09:14:28.180562 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70c2f8afa65a43ca9fe1af63181535c2a651816a5f44046db560f4716955f0cf" Mar 17 09:14:28 crc kubenswrapper[4813]: I0317 09:14:28.180645 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562314-dwd8b" Mar 17 09:14:28 crc kubenswrapper[4813]: I0317 09:14:28.192715 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92g4b\" (UniqueName: \"kubernetes.io/projected/9225f4aa-611e-447c-94b7-804e8c973203-kube-api-access-92g4b\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:28 crc kubenswrapper[4813]: I0317 09:14:28.261575 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 17 09:14:28 crc kubenswrapper[4813]: W0317 09:14:28.272560 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-podbbd05e02_97dc_48ce_b2d4_5d46d3ac63cb.slice/crio-3a9eefea65801d4686332bcfbc0c7e5258f233d282c09f9fa106f6e028c1b234 WatchSource:0}: Error finding container 3a9eefea65801d4686332bcfbc0c7e5258f233d282c09f9fa106f6e028c1b234: Status 404 returned error can't find the container with id 3a9eefea65801d4686332bcfbc0c7e5258f233d282c09f9fa106f6e028c1b234 Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.195518 4813 generic.go:334] "Generic (PLEG): container finished" podID="e49eaa7a-38bf-4996-ad34-64959cbccda2" containerID="4395a333d2cec6f2d56a2627f6cf0133dc3089cf05296bf74d047fc950f1d147" exitCode=0 Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.195910 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bf7m2" event={"ID":"e49eaa7a-38bf-4996-ad34-64959cbccda2","Type":"ContainerDied","Data":"4395a333d2cec6f2d56a2627f6cf0133dc3089cf05296bf74d047fc950f1d147"} Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.198228 4813 generic.go:334] "Generic (PLEG): container finished" podID="8626af43-6b5e-49e1-b19f-3986cdecdd2d" containerID="489850af96666df4ef1e82730a50eb36058eea72a391bebec30181a812172f63" exitCode=0 Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.198293 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562312-hjpcp" event={"ID":"8626af43-6b5e-49e1-b19f-3986cdecdd2d","Type":"ContainerDied","Data":"489850af96666df4ef1e82730a50eb36058eea72a391bebec30181a812172f63"} Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.206980 4813 generic.go:334] "Generic (PLEG): container finished" podID="6b3a014a-3574-4c2b-8331-df326d47f240" containerID="0f982775e6f3a4f0cea5af093812332c94e904833ea9524da65fc0bf13ce0552" exitCode=0 Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.207054 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6gvsx" event={"ID":"6b3a014a-3574-4c2b-8331-df326d47f240","Type":"ContainerDied","Data":"0f982775e6f3a4f0cea5af093812332c94e904833ea9524da65fc0bf13ce0552"} Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.210772 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-74f9cc6d4d-894cd"] Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.211041 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" podUID="7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0" containerName="controller-manager" containerID="cri-o://f226df3be2e759f1108da445677a823bda3cd97b637d30f9184454fbf25d23f6" gracePeriod=30 Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.224949 4813 generic.go:334] "Generic (PLEG): container finished" podID="bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb" containerID="c6215f595795456c42b09c8ce0f11b2b08a13cb26d174c6da12421853594d5ad" exitCode=0 Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.224996 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb","Type":"ContainerDied","Data":"c6215f595795456c42b09c8ce0f11b2b08a13cb26d174c6da12421853594d5ad"} Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.225021 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb","Type":"ContainerStarted","Data":"3a9eefea65801d4686332bcfbc0c7e5258f233d282c09f9fa106f6e028c1b234"} Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.315578 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb"] Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.315864 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" podUID="bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44" containerName="route-controller-manager" containerID="cri-o://a1ffec25ed32dda25b651929fec928c8bb12892cc5e8e71de10dc031edd0a35e" gracePeriod=30 Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.755813 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.815817 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.915174 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-client-ca\") pod \"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44\" (UID: \"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44\") " Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.915238 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-proxy-ca-bundles\") pod \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\" (UID: \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\") " Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.915271 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7kdd\" (UniqueName: \"kubernetes.io/projected/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-kube-api-access-m7kdd\") pod \"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44\" (UID: \"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44\") " Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.915316 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-serving-cert\") pod \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\" (UID: \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\") " Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.915347 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xfnzm\" (UniqueName: \"kubernetes.io/projected/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-kube-api-access-xfnzm\") pod \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\" (UID: \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\") " Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.915401 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-config\") pod \"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44\" (UID: \"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44\") " Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.915419 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-serving-cert\") pod \"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44\" (UID: \"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44\") " Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.915437 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-config\") pod \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\" (UID: \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\") " Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.915468 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-client-ca\") pod \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\" (UID: \"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0\") " Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.916045 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-client-ca" (OuterVolumeSpecName: "client-ca") pod "bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44" (UID: "bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.916059 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0" (UID: "7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.916064 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-client-ca" (OuterVolumeSpecName: "client-ca") pod "7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0" (UID: "7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.916580 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-config" (OuterVolumeSpecName: "config") pod "7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0" (UID: "7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.916967 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-config" (OuterVolumeSpecName: "config") pod "bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44" (UID: "bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.920044 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-kube-api-access-m7kdd" (OuterVolumeSpecName: "kube-api-access-m7kdd") pod "bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44" (UID: "bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44"). InnerVolumeSpecName "kube-api-access-m7kdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.920073 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0" (UID: "7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.920412 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-kube-api-access-xfnzm" (OuterVolumeSpecName: "kube-api-access-xfnzm") pod "7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0" (UID: "7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0"). InnerVolumeSpecName "kube-api-access-xfnzm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:14:29 crc kubenswrapper[4813]: I0317 09:14:29.921748 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44" (UID: "bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.017249 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.017274 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.017285 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.017293 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.017302 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.017310 4813 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.017320 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7kdd\" (UniqueName: \"kubernetes.io/projected/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44-kube-api-access-m7kdd\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.017329 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.017340 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xfnzm\" (UniqueName: \"kubernetes.io/projected/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0-kube-api-access-xfnzm\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.231238 4813 generic.go:334] "Generic (PLEG): container finished" podID="bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44" containerID="a1ffec25ed32dda25b651929fec928c8bb12892cc5e8e71de10dc031edd0a35e" exitCode=0 Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.231279 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.231320 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" event={"ID":"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44","Type":"ContainerDied","Data":"a1ffec25ed32dda25b651929fec928c8bb12892cc5e8e71de10dc031edd0a35e"} Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.231368 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb" event={"ID":"bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44","Type":"ContainerDied","Data":"cb570d829effa0a42fb5c2ed04213ea60556e7a604a01fa205c5ea2b73fa7ed5"} Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.231385 4813 scope.go:117] "RemoveContainer" containerID="a1ffec25ed32dda25b651929fec928c8bb12892cc5e8e71de10dc031edd0a35e" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.235596 4813 generic.go:334] "Generic (PLEG): container finished" podID="7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0" containerID="f226df3be2e759f1108da445677a823bda3cd97b637d30f9184454fbf25d23f6" exitCode=0 Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.235669 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" event={"ID":"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0","Type":"ContainerDied","Data":"f226df3be2e759f1108da445677a823bda3cd97b637d30f9184454fbf25d23f6"} Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.235677 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.235696 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74f9cc6d4d-894cd" event={"ID":"7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0","Type":"ContainerDied","Data":"7152088bb0ccf53b1a9d72144cecc214a6703ddf69872e47e8903c8ce412393c"} Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.238166 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6gvsx" event={"ID":"6b3a014a-3574-4c2b-8331-df326d47f240","Type":"ContainerStarted","Data":"3ccf2869040560c186883f68381aa6c02ceb23a455a78065f18335e2e2b78217"} Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.242540 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bf7m2" event={"ID":"e49eaa7a-38bf-4996-ad34-64959cbccda2","Type":"ContainerStarted","Data":"150c711cf5147379c0e341ae0163f834f4edcafe7ceea1557024f3432cfd2645"} Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.255790 4813 scope.go:117] "RemoveContainer" containerID="a1ffec25ed32dda25b651929fec928c8bb12892cc5e8e71de10dc031edd0a35e" Mar 17 09:14:30 crc kubenswrapper[4813]: E0317 09:14:30.256282 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1ffec25ed32dda25b651929fec928c8bb12892cc5e8e71de10dc031edd0a35e\": container with ID starting with a1ffec25ed32dda25b651929fec928c8bb12892cc5e8e71de10dc031edd0a35e not found: ID does not exist" containerID="a1ffec25ed32dda25b651929fec928c8bb12892cc5e8e71de10dc031edd0a35e" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.256311 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1ffec25ed32dda25b651929fec928c8bb12892cc5e8e71de10dc031edd0a35e"} err="failed to get container status \"a1ffec25ed32dda25b651929fec928c8bb12892cc5e8e71de10dc031edd0a35e\": rpc error: code = NotFound desc = could not find container \"a1ffec25ed32dda25b651929fec928c8bb12892cc5e8e71de10dc031edd0a35e\": container with ID starting with a1ffec25ed32dda25b651929fec928c8bb12892cc5e8e71de10dc031edd0a35e not found: ID does not exist" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.256334 4813 scope.go:117] "RemoveContainer" containerID="f226df3be2e759f1108da445677a823bda3cd97b637d30f9184454fbf25d23f6" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.267520 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6gvsx" podStartSLOduration=21.339221607 podStartE2EDuration="36.267502458s" podCreationTimestamp="2026-03-17 09:13:54 +0000 UTC" firstStartedPulling="2026-03-17 09:14:14.983277895 +0000 UTC m=+277.084081414" lastFinishedPulling="2026-03-17 09:14:29.911558766 +0000 UTC m=+292.012362265" observedRunningTime="2026-03-17 09:14:30.263510933 +0000 UTC m=+292.364314442" watchObservedRunningTime="2026-03-17 09:14:30.267502458 +0000 UTC m=+292.368305957" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.286867 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-74f9cc6d4d-894cd"] Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.291591 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-74f9cc6d4d-894cd"] Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.305316 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-bf7m2" podStartSLOduration=26.488652231 podStartE2EDuration="36.305293826s" podCreationTimestamp="2026-03-17 09:13:54 +0000 UTC" firstStartedPulling="2026-03-17 09:14:20.050622363 +0000 UTC m=+282.151425862" lastFinishedPulling="2026-03-17 09:14:29.867263958 +0000 UTC m=+291.968067457" observedRunningTime="2026-03-17 09:14:30.303739192 +0000 UTC m=+292.404542691" watchObservedRunningTime="2026-03-17 09:14:30.305293826 +0000 UTC m=+292.406097345" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.310435 4813 scope.go:117] "RemoveContainer" containerID="f226df3be2e759f1108da445677a823bda3cd97b637d30f9184454fbf25d23f6" Mar 17 09:14:30 crc kubenswrapper[4813]: E0317 09:14:30.310938 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f226df3be2e759f1108da445677a823bda3cd97b637d30f9184454fbf25d23f6\": container with ID starting with f226df3be2e759f1108da445677a823bda3cd97b637d30f9184454fbf25d23f6 not found: ID does not exist" containerID="f226df3be2e759f1108da445677a823bda3cd97b637d30f9184454fbf25d23f6" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.311045 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f226df3be2e759f1108da445677a823bda3cd97b637d30f9184454fbf25d23f6"} err="failed to get container status \"f226df3be2e759f1108da445677a823bda3cd97b637d30f9184454fbf25d23f6\": rpc error: code = NotFound desc = could not find container \"f226df3be2e759f1108da445677a823bda3cd97b637d30f9184454fbf25d23f6\": container with ID starting with f226df3be2e759f1108da445677a823bda3cd97b637d30f9184454fbf25d23f6 not found: ID does not exist" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.320932 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb"] Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.324735 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-647b8ff8cf-w5vvb"] Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.514236 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562312-hjpcp" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.517188 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.625005 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcghd\" (UniqueName: \"kubernetes.io/projected/8626af43-6b5e-49e1-b19f-3986cdecdd2d-kube-api-access-fcghd\") pod \"8626af43-6b5e-49e1-b19f-3986cdecdd2d\" (UID: \"8626af43-6b5e-49e1-b19f-3986cdecdd2d\") " Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.625287 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb-kubelet-dir\") pod \"bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb\" (UID: \"bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb\") " Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.625317 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb-kube-api-access\") pod \"bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb\" (UID: \"bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb\") " Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.625719 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb" (UID: "bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.630072 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb" (UID: "bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.630308 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8626af43-6b5e-49e1-b19f-3986cdecdd2d-kube-api-access-fcghd" (OuterVolumeSpecName: "kube-api-access-fcghd") pod "8626af43-6b5e-49e1-b19f-3986cdecdd2d" (UID: "8626af43-6b5e-49e1-b19f-3986cdecdd2d"). InnerVolumeSpecName "kube-api-access-fcghd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.727045 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcghd\" (UniqueName: \"kubernetes.io/projected/8626af43-6b5e-49e1-b19f-3986cdecdd2d-kube-api-access-fcghd\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.727362 4813 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.727512 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.739174 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0" path="/var/lib/kubelet/pods/7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0/volumes" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.739931 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44" path="/var/lib/kubelet/pods/bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44/volumes" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.935304 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz"] Mar 17 09:14:30 crc kubenswrapper[4813]: E0317 09:14:30.935627 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0" containerName="controller-manager" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.935647 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0" containerName="controller-manager" Mar 17 09:14:30 crc kubenswrapper[4813]: E0317 09:14:30.935664 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44" containerName="route-controller-manager" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.935673 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44" containerName="route-controller-manager" Mar 17 09:14:30 crc kubenswrapper[4813]: E0317 09:14:30.935683 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8626af43-6b5e-49e1-b19f-3986cdecdd2d" containerName="oc" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.935691 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8626af43-6b5e-49e1-b19f-3986cdecdd2d" containerName="oc" Mar 17 09:14:30 crc kubenswrapper[4813]: E0317 09:14:30.935702 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb" containerName="pruner" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.935710 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb" containerName="pruner" Mar 17 09:14:30 crc kubenswrapper[4813]: E0317 09:14:30.935728 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9225f4aa-611e-447c-94b7-804e8c973203" containerName="oc" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.935737 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9225f4aa-611e-447c-94b7-804e8c973203" containerName="oc" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.935867 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d000133-b0b4-43c8-b7b5-6f3b6de1a3f0" containerName="controller-manager" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.935881 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9225f4aa-611e-447c-94b7-804e8c973203" containerName="oc" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.935891 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc9de28e-cda4-4f04-a6c1-64c7ee2d2e44" containerName="route-controller-manager" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.935902 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb" containerName="pruner" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.935917 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8626af43-6b5e-49e1-b19f-3986cdecdd2d" containerName="oc" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.936325 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.936878 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd"] Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.937500 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.938312 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.939066 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.940744 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.940782 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.941121 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.941377 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.941097 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.941845 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.942075 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.942366 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.942487 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.942564 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.946362 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz"] Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.947215 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 17 09:14:30 crc kubenswrapper[4813]: I0317 09:14:30.960395 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd"] Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.032047 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/788a513c-9124-4152-908c-df0c059b6270-proxy-ca-bundles\") pod \"controller-manager-7d655f7fcd-rrwmd\" (UID: \"788a513c-9124-4152-908c-df0c059b6270\") " pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.032144 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvqkb\" (UniqueName: \"kubernetes.io/projected/788a513c-9124-4152-908c-df0c059b6270-kube-api-access-vvqkb\") pod \"controller-manager-7d655f7fcd-rrwmd\" (UID: \"788a513c-9124-4152-908c-df0c059b6270\") " pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.032193 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/788a513c-9124-4152-908c-df0c059b6270-config\") pod \"controller-manager-7d655f7fcd-rrwmd\" (UID: \"788a513c-9124-4152-908c-df0c059b6270\") " pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.032241 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28qpw\" (UniqueName: \"kubernetes.io/projected/48a312b3-3ba5-40ab-bbc9-b4b21903f978-kube-api-access-28qpw\") pod \"route-controller-manager-fd576fb67-mvgcz\" (UID: \"48a312b3-3ba5-40ab-bbc9-b4b21903f978\") " pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.032318 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/788a513c-9124-4152-908c-df0c059b6270-serving-cert\") pod \"controller-manager-7d655f7fcd-rrwmd\" (UID: \"788a513c-9124-4152-908c-df0c059b6270\") " pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.032385 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/48a312b3-3ba5-40ab-bbc9-b4b21903f978-client-ca\") pod \"route-controller-manager-fd576fb67-mvgcz\" (UID: \"48a312b3-3ba5-40ab-bbc9-b4b21903f978\") " pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.032455 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48a312b3-3ba5-40ab-bbc9-b4b21903f978-config\") pod \"route-controller-manager-fd576fb67-mvgcz\" (UID: \"48a312b3-3ba5-40ab-bbc9-b4b21903f978\") " pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.032480 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48a312b3-3ba5-40ab-bbc9-b4b21903f978-serving-cert\") pod \"route-controller-manager-fd576fb67-mvgcz\" (UID: \"48a312b3-3ba5-40ab-bbc9-b4b21903f978\") " pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.032505 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/788a513c-9124-4152-908c-df0c059b6270-client-ca\") pod \"controller-manager-7d655f7fcd-rrwmd\" (UID: \"788a513c-9124-4152-908c-df0c059b6270\") " pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.077803 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.078555 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.081902 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.133967 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48a312b3-3ba5-40ab-bbc9-b4b21903f978-config\") pod \"route-controller-manager-fd576fb67-mvgcz\" (UID: \"48a312b3-3ba5-40ab-bbc9-b4b21903f978\") " pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.134019 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48a312b3-3ba5-40ab-bbc9-b4b21903f978-serving-cert\") pod \"route-controller-manager-fd576fb67-mvgcz\" (UID: \"48a312b3-3ba5-40ab-bbc9-b4b21903f978\") " pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.134049 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/788a513c-9124-4152-908c-df0c059b6270-client-ca\") pod \"controller-manager-7d655f7fcd-rrwmd\" (UID: \"788a513c-9124-4152-908c-df0c059b6270\") " pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.134077 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/788a513c-9124-4152-908c-df0c059b6270-proxy-ca-bundles\") pod \"controller-manager-7d655f7fcd-rrwmd\" (UID: \"788a513c-9124-4152-908c-df0c059b6270\") " pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.134098 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvqkb\" (UniqueName: \"kubernetes.io/projected/788a513c-9124-4152-908c-df0c059b6270-kube-api-access-vvqkb\") pod \"controller-manager-7d655f7fcd-rrwmd\" (UID: \"788a513c-9124-4152-908c-df0c059b6270\") " pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.134121 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/788a513c-9124-4152-908c-df0c059b6270-config\") pod \"controller-manager-7d655f7fcd-rrwmd\" (UID: \"788a513c-9124-4152-908c-df0c059b6270\") " pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.134150 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28qpw\" (UniqueName: \"kubernetes.io/projected/48a312b3-3ba5-40ab-bbc9-b4b21903f978-kube-api-access-28qpw\") pod \"route-controller-manager-fd576fb67-mvgcz\" (UID: \"48a312b3-3ba5-40ab-bbc9-b4b21903f978\") " pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.134181 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/788a513c-9124-4152-908c-df0c059b6270-serving-cert\") pod \"controller-manager-7d655f7fcd-rrwmd\" (UID: \"788a513c-9124-4152-908c-df0c059b6270\") " pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.134231 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/48a312b3-3ba5-40ab-bbc9-b4b21903f978-client-ca\") pod \"route-controller-manager-fd576fb67-mvgcz\" (UID: \"48a312b3-3ba5-40ab-bbc9-b4b21903f978\") " pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.135352 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/48a312b3-3ba5-40ab-bbc9-b4b21903f978-client-ca\") pod \"route-controller-manager-fd576fb67-mvgcz\" (UID: \"48a312b3-3ba5-40ab-bbc9-b4b21903f978\") " pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.135394 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/788a513c-9124-4152-908c-df0c059b6270-client-ca\") pod \"controller-manager-7d655f7fcd-rrwmd\" (UID: \"788a513c-9124-4152-908c-df0c059b6270\") " pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.135715 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/788a513c-9124-4152-908c-df0c059b6270-config\") pod \"controller-manager-7d655f7fcd-rrwmd\" (UID: \"788a513c-9124-4152-908c-df0c059b6270\") " pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.135728 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48a312b3-3ba5-40ab-bbc9-b4b21903f978-config\") pod \"route-controller-manager-fd576fb67-mvgcz\" (UID: \"48a312b3-3ba5-40ab-bbc9-b4b21903f978\") " pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.136198 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/788a513c-9124-4152-908c-df0c059b6270-proxy-ca-bundles\") pod \"controller-manager-7d655f7fcd-rrwmd\" (UID: \"788a513c-9124-4152-908c-df0c059b6270\") " pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.137576 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/788a513c-9124-4152-908c-df0c059b6270-serving-cert\") pod \"controller-manager-7d655f7fcd-rrwmd\" (UID: \"788a513c-9124-4152-908c-df0c059b6270\") " pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.138609 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48a312b3-3ba5-40ab-bbc9-b4b21903f978-serving-cert\") pod \"route-controller-manager-fd576fb67-mvgcz\" (UID: \"48a312b3-3ba5-40ab-bbc9-b4b21903f978\") " pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.153496 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvqkb\" (UniqueName: \"kubernetes.io/projected/788a513c-9124-4152-908c-df0c059b6270-kube-api-access-vvqkb\") pod \"controller-manager-7d655f7fcd-rrwmd\" (UID: \"788a513c-9124-4152-908c-df0c059b6270\") " pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.153526 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28qpw\" (UniqueName: \"kubernetes.io/projected/48a312b3-3ba5-40ab-bbc9-b4b21903f978-kube-api-access-28qpw\") pod \"route-controller-manager-fd576fb67-mvgcz\" (UID: \"48a312b3-3ba5-40ab-bbc9-b4b21903f978\") " pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.235657 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e87d1b34-2547-4acd-8843-a8bffc737563-kubelet-dir\") pod \"installer-9-crc\" (UID: \"e87d1b34-2547-4acd-8843-a8bffc737563\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.236825 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e87d1b34-2547-4acd-8843-a8bffc737563-var-lock\") pod \"installer-9-crc\" (UID: \"e87d1b34-2547-4acd-8843-a8bffc737563\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.236948 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e87d1b34-2547-4acd-8843-a8bffc737563-kube-api-access\") pod \"installer-9-crc\" (UID: \"e87d1b34-2547-4acd-8843-a8bffc737563\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.249807 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"bbd05e02-97dc-48ce-b2d4-5d46d3ac63cb","Type":"ContainerDied","Data":"3a9eefea65801d4686332bcfbc0c7e5258f233d282c09f9fa106f6e028c1b234"} Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.250020 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3a9eefea65801d4686332bcfbc0c7e5258f233d282c09f9fa106f6e028c1b234" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.249818 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.251893 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562312-hjpcp" event={"ID":"8626af43-6b5e-49e1-b19f-3986cdecdd2d","Type":"ContainerDied","Data":"13518e37a85cf60fdf0ccedf00537690826be4b4ec70bbb11418dd03a067aa1c"} Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.251926 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="13518e37a85cf60fdf0ccedf00537690826be4b4ec70bbb11418dd03a067aa1c" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.251960 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562312-hjpcp" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.260277 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.275618 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.337834 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e87d1b34-2547-4acd-8843-a8bffc737563-kubelet-dir\") pod \"installer-9-crc\" (UID: \"e87d1b34-2547-4acd-8843-a8bffc737563\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.337989 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e87d1b34-2547-4acd-8843-a8bffc737563-kubelet-dir\") pod \"installer-9-crc\" (UID: \"e87d1b34-2547-4acd-8843-a8bffc737563\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.338013 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e87d1b34-2547-4acd-8843-a8bffc737563-var-lock\") pod \"installer-9-crc\" (UID: \"e87d1b34-2547-4acd-8843-a8bffc737563\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.338033 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e87d1b34-2547-4acd-8843-a8bffc737563-kube-api-access\") pod \"installer-9-crc\" (UID: \"e87d1b34-2547-4acd-8843-a8bffc737563\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.338058 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e87d1b34-2547-4acd-8843-a8bffc737563-var-lock\") pod \"installer-9-crc\" (UID: \"e87d1b34-2547-4acd-8843-a8bffc737563\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.358328 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e87d1b34-2547-4acd-8843-a8bffc737563-kube-api-access\") pod \"installer-9-crc\" (UID: \"e87d1b34-2547-4acd-8843-a8bffc737563\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.395087 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.522291 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2skjk" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.565210 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd"] Mar 17 09:14:31 crc kubenswrapper[4813]: W0317 09:14:31.590200 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod788a513c_9124_4152_908c_df0c059b6270.slice/crio-80da5c48cabb45575212074b1db4381bb00fd081cf4ef39fb6c96c3fa2b583ab WatchSource:0}: Error finding container 80da5c48cabb45575212074b1db4381bb00fd081cf4ef39fb6c96c3fa2b583ab: Status 404 returned error can't find the container with id 80da5c48cabb45575212074b1db4381bb00fd081cf4ef39fb6c96c3fa2b583ab Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.593364 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2skjk" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.668076 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.671494 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz"] Mar 17 09:14:31 crc kubenswrapper[4813]: W0317 09:14:31.692701 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pode87d1b34_2547_4acd_8843_a8bffc737563.slice/crio-e4bf54f7cff4843be1351bd95b6e5fd183e2b1043c2434596bd48b41348281ae WatchSource:0}: Error finding container e4bf54f7cff4843be1351bd95b6e5fd183e2b1043c2434596bd48b41348281ae: Status 404 returned error can't find the container with id e4bf54f7cff4843be1351bd95b6e5fd183e2b1043c2434596bd48b41348281ae Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.926865 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5dpjf" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.962244 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 17 09:14:31 crc kubenswrapper[4813]: I0317 09:14:31.972337 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5dpjf" Mar 17 09:14:32 crc kubenswrapper[4813]: I0317 09:14:32.265217 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" event={"ID":"48a312b3-3ba5-40ab-bbc9-b4b21903f978","Type":"ContainerStarted","Data":"aa7b6baf3551238cdeea1f20748a58f0b60f5a023e3f01815149a05c4fe21c1d"} Mar 17 09:14:32 crc kubenswrapper[4813]: I0317 09:14:32.265651 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" event={"ID":"48a312b3-3ba5-40ab-bbc9-b4b21903f978","Type":"ContainerStarted","Data":"2363b0c6a8f8f147e80ec520f30457a83fcd62b102dc4e4dd084b7f7d91ec6e1"} Mar 17 09:14:32 crc kubenswrapper[4813]: I0317 09:14:32.265679 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" Mar 17 09:14:32 crc kubenswrapper[4813]: I0317 09:14:32.266297 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e87d1b34-2547-4acd-8843-a8bffc737563","Type":"ContainerStarted","Data":"31bded1816cfd4a21a18bdf8b38e4e14c4cc7e06adde9dc74da11540e79e2499"} Mar 17 09:14:32 crc kubenswrapper[4813]: I0317 09:14:32.266339 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e87d1b34-2547-4acd-8843-a8bffc737563","Type":"ContainerStarted","Data":"e4bf54f7cff4843be1351bd95b6e5fd183e2b1043c2434596bd48b41348281ae"} Mar 17 09:14:32 crc kubenswrapper[4813]: I0317 09:14:32.267496 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" event={"ID":"788a513c-9124-4152-908c-df0c059b6270","Type":"ContainerStarted","Data":"346f1361a81b5452be41bbe4ac74ab25262853f7c34d73b5f1b23d272f77ade6"} Mar 17 09:14:32 crc kubenswrapper[4813]: I0317 09:14:32.267527 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" event={"ID":"788a513c-9124-4152-908c-df0c059b6270","Type":"ContainerStarted","Data":"80da5c48cabb45575212074b1db4381bb00fd081cf4ef39fb6c96c3fa2b583ab"} Mar 17 09:14:32 crc kubenswrapper[4813]: I0317 09:14:32.267890 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" Mar 17 09:14:32 crc kubenswrapper[4813]: I0317 09:14:32.273152 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" Mar 17 09:14:32 crc kubenswrapper[4813]: I0317 09:14:32.293529 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" podStartSLOduration=3.293511924 podStartE2EDuration="3.293511924s" podCreationTimestamp="2026-03-17 09:14:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:14:32.290894965 +0000 UTC m=+294.391698464" watchObservedRunningTime="2026-03-17 09:14:32.293511924 +0000 UTC m=+294.394315423" Mar 17 09:14:32 crc kubenswrapper[4813]: I0317 09:14:32.313324 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" podStartSLOduration=3.313295368 podStartE2EDuration="3.313295368s" podCreationTimestamp="2026-03-17 09:14:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:14:32.311851619 +0000 UTC m=+294.412655118" watchObservedRunningTime="2026-03-17 09:14:32.313295368 +0000 UTC m=+294.414098867" Mar 17 09:14:32 crc kubenswrapper[4813]: I0317 09:14:32.340834 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=1.3408166750000001 podStartE2EDuration="1.340816675s" podCreationTimestamp="2026-03-17 09:14:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:14:32.338788796 +0000 UTC m=+294.439592285" watchObservedRunningTime="2026-03-17 09:14:32.340816675 +0000 UTC m=+294.441620174" Mar 17 09:14:32 crc kubenswrapper[4813]: I0317 09:14:32.407904 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" Mar 17 09:14:32 crc kubenswrapper[4813]: I0317 09:14:32.888069 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5kp7w"] Mar 17 09:14:33 crc kubenswrapper[4813]: I0317 09:14:33.176968 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5dpjf"] Mar 17 09:14:33 crc kubenswrapper[4813]: I0317 09:14:33.272586 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5dpjf" podUID="b7d1b948-0f9a-4ba3-9f58-88590e13ba63" containerName="registry-server" containerID="cri-o://97ebdfe561e724beec2af09e7bff0a01b83678538da9176586cefd595e83d7ca" gracePeriod=2 Mar 17 09:14:33 crc kubenswrapper[4813]: I0317 09:14:33.544866 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-js422" Mar 17 09:14:33 crc kubenswrapper[4813]: I0317 09:14:33.683776 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5dpjf" Mar 17 09:14:33 crc kubenswrapper[4813]: I0317 09:14:33.778788 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5vxln\" (UniqueName: \"kubernetes.io/projected/b7d1b948-0f9a-4ba3-9f58-88590e13ba63-kube-api-access-5vxln\") pod \"b7d1b948-0f9a-4ba3-9f58-88590e13ba63\" (UID: \"b7d1b948-0f9a-4ba3-9f58-88590e13ba63\") " Mar 17 09:14:33 crc kubenswrapper[4813]: I0317 09:14:33.778830 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7d1b948-0f9a-4ba3-9f58-88590e13ba63-utilities\") pod \"b7d1b948-0f9a-4ba3-9f58-88590e13ba63\" (UID: \"b7d1b948-0f9a-4ba3-9f58-88590e13ba63\") " Mar 17 09:14:33 crc kubenswrapper[4813]: I0317 09:14:33.778878 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7d1b948-0f9a-4ba3-9f58-88590e13ba63-catalog-content\") pod \"b7d1b948-0f9a-4ba3-9f58-88590e13ba63\" (UID: \"b7d1b948-0f9a-4ba3-9f58-88590e13ba63\") " Mar 17 09:14:33 crc kubenswrapper[4813]: I0317 09:14:33.785800 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7d1b948-0f9a-4ba3-9f58-88590e13ba63-utilities" (OuterVolumeSpecName: "utilities") pod "b7d1b948-0f9a-4ba3-9f58-88590e13ba63" (UID: "b7d1b948-0f9a-4ba3-9f58-88590e13ba63"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:14:33 crc kubenswrapper[4813]: I0317 09:14:33.790471 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7d1b948-0f9a-4ba3-9f58-88590e13ba63-kube-api-access-5vxln" (OuterVolumeSpecName: "kube-api-access-5vxln") pod "b7d1b948-0f9a-4ba3-9f58-88590e13ba63" (UID: "b7d1b948-0f9a-4ba3-9f58-88590e13ba63"). InnerVolumeSpecName "kube-api-access-5vxln". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:14:33 crc kubenswrapper[4813]: I0317 09:14:33.833807 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7d1b948-0f9a-4ba3-9f58-88590e13ba63-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b7d1b948-0f9a-4ba3-9f58-88590e13ba63" (UID: "b7d1b948-0f9a-4ba3-9f58-88590e13ba63"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:14:33 crc kubenswrapper[4813]: I0317 09:14:33.880275 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5vxln\" (UniqueName: \"kubernetes.io/projected/b7d1b948-0f9a-4ba3-9f58-88590e13ba63-kube-api-access-5vxln\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:33 crc kubenswrapper[4813]: I0317 09:14:33.880316 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7d1b948-0f9a-4ba3-9f58-88590e13ba63-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:33 crc kubenswrapper[4813]: I0317 09:14:33.880352 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7d1b948-0f9a-4ba3-9f58-88590e13ba63-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:33 crc kubenswrapper[4813]: I0317 09:14:33.944874 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-x95rg" Mar 17 09:14:34 crc kubenswrapper[4813]: I0317 09:14:34.279806 4813 generic.go:334] "Generic (PLEG): container finished" podID="b7d1b948-0f9a-4ba3-9f58-88590e13ba63" containerID="97ebdfe561e724beec2af09e7bff0a01b83678538da9176586cefd595e83d7ca" exitCode=0 Mar 17 09:14:34 crc kubenswrapper[4813]: I0317 09:14:34.279866 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5dpjf" Mar 17 09:14:34 crc kubenswrapper[4813]: I0317 09:14:34.279897 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5dpjf" event={"ID":"b7d1b948-0f9a-4ba3-9f58-88590e13ba63","Type":"ContainerDied","Data":"97ebdfe561e724beec2af09e7bff0a01b83678538da9176586cefd595e83d7ca"} Mar 17 09:14:34 crc kubenswrapper[4813]: I0317 09:14:34.280288 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5dpjf" event={"ID":"b7d1b948-0f9a-4ba3-9f58-88590e13ba63","Type":"ContainerDied","Data":"d93323d902b018edef2b9ab9bdf52561a8da84a2a7057d1376b6383b00a51846"} Mar 17 09:14:34 crc kubenswrapper[4813]: I0317 09:14:34.280319 4813 scope.go:117] "RemoveContainer" containerID="97ebdfe561e724beec2af09e7bff0a01b83678538da9176586cefd595e83d7ca" Mar 17 09:14:34 crc kubenswrapper[4813]: I0317 09:14:34.298842 4813 scope.go:117] "RemoveContainer" containerID="2feaf20e554846b8e9aa047cc5fe61f1c3a7b837e0fc677962aee650e02877b8" Mar 17 09:14:34 crc kubenswrapper[4813]: I0317 09:14:34.309135 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5dpjf"] Mar 17 09:14:34 crc kubenswrapper[4813]: I0317 09:14:34.315244 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5dpjf"] Mar 17 09:14:34 crc kubenswrapper[4813]: I0317 09:14:34.326495 4813 scope.go:117] "RemoveContainer" containerID="a6952169c00766088a53864389b265cec117256844467b5537f5d7d067ec0459" Mar 17 09:14:34 crc kubenswrapper[4813]: I0317 09:14:34.340623 4813 scope.go:117] "RemoveContainer" containerID="97ebdfe561e724beec2af09e7bff0a01b83678538da9176586cefd595e83d7ca" Mar 17 09:14:34 crc kubenswrapper[4813]: E0317 09:14:34.341159 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"97ebdfe561e724beec2af09e7bff0a01b83678538da9176586cefd595e83d7ca\": container with ID starting with 97ebdfe561e724beec2af09e7bff0a01b83678538da9176586cefd595e83d7ca not found: ID does not exist" containerID="97ebdfe561e724beec2af09e7bff0a01b83678538da9176586cefd595e83d7ca" Mar 17 09:14:34 crc kubenswrapper[4813]: I0317 09:14:34.341202 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"97ebdfe561e724beec2af09e7bff0a01b83678538da9176586cefd595e83d7ca"} err="failed to get container status \"97ebdfe561e724beec2af09e7bff0a01b83678538da9176586cefd595e83d7ca\": rpc error: code = NotFound desc = could not find container \"97ebdfe561e724beec2af09e7bff0a01b83678538da9176586cefd595e83d7ca\": container with ID starting with 97ebdfe561e724beec2af09e7bff0a01b83678538da9176586cefd595e83d7ca not found: ID does not exist" Mar 17 09:14:34 crc kubenswrapper[4813]: I0317 09:14:34.341237 4813 scope.go:117] "RemoveContainer" containerID="2feaf20e554846b8e9aa047cc5fe61f1c3a7b837e0fc677962aee650e02877b8" Mar 17 09:14:34 crc kubenswrapper[4813]: E0317 09:14:34.341558 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2feaf20e554846b8e9aa047cc5fe61f1c3a7b837e0fc677962aee650e02877b8\": container with ID starting with 2feaf20e554846b8e9aa047cc5fe61f1c3a7b837e0fc677962aee650e02877b8 not found: ID does not exist" containerID="2feaf20e554846b8e9aa047cc5fe61f1c3a7b837e0fc677962aee650e02877b8" Mar 17 09:14:34 crc kubenswrapper[4813]: I0317 09:14:34.341581 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2feaf20e554846b8e9aa047cc5fe61f1c3a7b837e0fc677962aee650e02877b8"} err="failed to get container status \"2feaf20e554846b8e9aa047cc5fe61f1c3a7b837e0fc677962aee650e02877b8\": rpc error: code = NotFound desc = could not find container \"2feaf20e554846b8e9aa047cc5fe61f1c3a7b837e0fc677962aee650e02877b8\": container with ID starting with 2feaf20e554846b8e9aa047cc5fe61f1c3a7b837e0fc677962aee650e02877b8 not found: ID does not exist" Mar 17 09:14:34 crc kubenswrapper[4813]: I0317 09:14:34.341610 4813 scope.go:117] "RemoveContainer" containerID="a6952169c00766088a53864389b265cec117256844467b5537f5d7d067ec0459" Mar 17 09:14:34 crc kubenswrapper[4813]: E0317 09:14:34.341885 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6952169c00766088a53864389b265cec117256844467b5537f5d7d067ec0459\": container with ID starting with a6952169c00766088a53864389b265cec117256844467b5537f5d7d067ec0459 not found: ID does not exist" containerID="a6952169c00766088a53864389b265cec117256844467b5537f5d7d067ec0459" Mar 17 09:14:34 crc kubenswrapper[4813]: I0317 09:14:34.341923 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6952169c00766088a53864389b265cec117256844467b5537f5d7d067ec0459"} err="failed to get container status \"a6952169c00766088a53864389b265cec117256844467b5537f5d7d067ec0459\": rpc error: code = NotFound desc = could not find container \"a6952169c00766088a53864389b265cec117256844467b5537f5d7d067ec0459\": container with ID starting with a6952169c00766088a53864389b265cec117256844467b5537f5d7d067ec0459 not found: ID does not exist" Mar 17 09:14:34 crc kubenswrapper[4813]: I0317 09:14:34.693164 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6gvsx" Mar 17 09:14:34 crc kubenswrapper[4813]: I0317 09:14:34.693227 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6gvsx" Mar 17 09:14:34 crc kubenswrapper[4813]: I0317 09:14:34.741430 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7d1b948-0f9a-4ba3-9f58-88590e13ba63" path="/var/lib/kubelet/pods/b7d1b948-0f9a-4ba3-9f58-88590e13ba63/volumes" Mar 17 09:14:35 crc kubenswrapper[4813]: I0317 09:14:35.109482 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-bf7m2" Mar 17 09:14:35 crc kubenswrapper[4813]: I0317 09:14:35.109546 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-bf7m2" Mar 17 09:14:35 crc kubenswrapper[4813]: I0317 09:14:35.733980 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-6gvsx" podUID="6b3a014a-3574-4c2b-8331-df326d47f240" containerName="registry-server" probeResult="failure" output=< Mar 17 09:14:35 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Mar 17 09:14:35 crc kubenswrapper[4813]: > Mar 17 09:14:36 crc kubenswrapper[4813]: I0317 09:14:36.160121 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-bf7m2" podUID="e49eaa7a-38bf-4996-ad34-64959cbccda2" containerName="registry-server" probeResult="failure" output=< Mar 17 09:14:36 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Mar 17 09:14:36 crc kubenswrapper[4813]: > Mar 17 09:14:36 crc kubenswrapper[4813]: I0317 09:14:36.294957 4813 generic.go:334] "Generic (PLEG): container finished" podID="404fc67e-83be-4be6-bf6e-4279c0543316" containerID="2097079cc5e14a78c96d867fbb7b8d994d7f551495e284af871ea520ff284874" exitCode=0 Mar 17 09:14:36 crc kubenswrapper[4813]: I0317 09:14:36.295032 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5lhdz" event={"ID":"404fc67e-83be-4be6-bf6e-4279c0543316","Type":"ContainerDied","Data":"2097079cc5e14a78c96d867fbb7b8d994d7f551495e284af871ea520ff284874"} Mar 17 09:14:36 crc kubenswrapper[4813]: I0317 09:14:36.296928 4813 generic.go:334] "Generic (PLEG): container finished" podID="1b9e1de0-da0f-422e-a79b-4fb54d2e3acb" containerID="4d58dd63365d261f37b02064cbccd13904fbc1ce79697a65ff95c2fee7013b3e" exitCode=0 Mar 17 09:14:36 crc kubenswrapper[4813]: I0317 09:14:36.296955 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxrc7" event={"ID":"1b9e1de0-da0f-422e-a79b-4fb54d2e3acb","Type":"ContainerDied","Data":"4d58dd63365d261f37b02064cbccd13904fbc1ce79697a65ff95c2fee7013b3e"} Mar 17 09:14:36 crc kubenswrapper[4813]: I0317 09:14:36.972916 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x95rg"] Mar 17 09:14:36 crc kubenswrapper[4813]: I0317 09:14:36.973496 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-x95rg" podUID="4ca1cd05-e8fc-47b7-904b-a0a544f654b0" containerName="registry-server" containerID="cri-o://988ee7f8b66efd6561f1e628375cfe7016af93e9579b928a9f95754744bb9ac7" gracePeriod=2 Mar 17 09:14:37 crc kubenswrapper[4813]: I0317 09:14:37.315917 4813 generic.go:334] "Generic (PLEG): container finished" podID="4ca1cd05-e8fc-47b7-904b-a0a544f654b0" containerID="988ee7f8b66efd6561f1e628375cfe7016af93e9579b928a9f95754744bb9ac7" exitCode=0 Mar 17 09:14:37 crc kubenswrapper[4813]: I0317 09:14:37.315983 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x95rg" event={"ID":"4ca1cd05-e8fc-47b7-904b-a0a544f654b0","Type":"ContainerDied","Data":"988ee7f8b66efd6561f1e628375cfe7016af93e9579b928a9f95754744bb9ac7"} Mar 17 09:14:37 crc kubenswrapper[4813]: I0317 09:14:37.320500 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxrc7" event={"ID":"1b9e1de0-da0f-422e-a79b-4fb54d2e3acb","Type":"ContainerStarted","Data":"76af5a3320222301f14aee13eb616c8cc0bfc67f5aca3d2f0701ca64340366b2"} Mar 17 09:14:37 crc kubenswrapper[4813]: I0317 09:14:37.342182 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-lxrc7" podStartSLOduration=2.118414944 podStartE2EDuration="46.342165126s" podCreationTimestamp="2026-03-17 09:13:51 +0000 UTC" firstStartedPulling="2026-03-17 09:13:52.559065836 +0000 UTC m=+254.659869335" lastFinishedPulling="2026-03-17 09:14:36.782816018 +0000 UTC m=+298.883619517" observedRunningTime="2026-03-17 09:14:37.339788905 +0000 UTC m=+299.440592424" watchObservedRunningTime="2026-03-17 09:14:37.342165126 +0000 UTC m=+299.442968625" Mar 17 09:14:37 crc kubenswrapper[4813]: I0317 09:14:37.412490 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x95rg" Mar 17 09:14:37 crc kubenswrapper[4813]: I0317 09:14:37.526563 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ca1cd05-e8fc-47b7-904b-a0a544f654b0-utilities\") pod \"4ca1cd05-e8fc-47b7-904b-a0a544f654b0\" (UID: \"4ca1cd05-e8fc-47b7-904b-a0a544f654b0\") " Mar 17 09:14:37 crc kubenswrapper[4813]: I0317 09:14:37.526651 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6z7ks\" (UniqueName: \"kubernetes.io/projected/4ca1cd05-e8fc-47b7-904b-a0a544f654b0-kube-api-access-6z7ks\") pod \"4ca1cd05-e8fc-47b7-904b-a0a544f654b0\" (UID: \"4ca1cd05-e8fc-47b7-904b-a0a544f654b0\") " Mar 17 09:14:37 crc kubenswrapper[4813]: I0317 09:14:37.526726 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ca1cd05-e8fc-47b7-904b-a0a544f654b0-catalog-content\") pod \"4ca1cd05-e8fc-47b7-904b-a0a544f654b0\" (UID: \"4ca1cd05-e8fc-47b7-904b-a0a544f654b0\") " Mar 17 09:14:37 crc kubenswrapper[4813]: I0317 09:14:37.527229 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ca1cd05-e8fc-47b7-904b-a0a544f654b0-utilities" (OuterVolumeSpecName: "utilities") pod "4ca1cd05-e8fc-47b7-904b-a0a544f654b0" (UID: "4ca1cd05-e8fc-47b7-904b-a0a544f654b0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:14:37 crc kubenswrapper[4813]: I0317 09:14:37.534240 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ca1cd05-e8fc-47b7-904b-a0a544f654b0-kube-api-access-6z7ks" (OuterVolumeSpecName: "kube-api-access-6z7ks") pod "4ca1cd05-e8fc-47b7-904b-a0a544f654b0" (UID: "4ca1cd05-e8fc-47b7-904b-a0a544f654b0"). InnerVolumeSpecName "kube-api-access-6z7ks". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:14:37 crc kubenswrapper[4813]: I0317 09:14:37.554412 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4ca1cd05-e8fc-47b7-904b-a0a544f654b0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4ca1cd05-e8fc-47b7-904b-a0a544f654b0" (UID: "4ca1cd05-e8fc-47b7-904b-a0a544f654b0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:14:37 crc kubenswrapper[4813]: I0317 09:14:37.628407 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ca1cd05-e8fc-47b7-904b-a0a544f654b0-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:37 crc kubenswrapper[4813]: I0317 09:14:37.628446 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ca1cd05-e8fc-47b7-904b-a0a544f654b0-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:37 crc kubenswrapper[4813]: I0317 09:14:37.628459 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6z7ks\" (UniqueName: \"kubernetes.io/projected/4ca1cd05-e8fc-47b7-904b-a0a544f654b0-kube-api-access-6z7ks\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:38 crc kubenswrapper[4813]: I0317 09:14:38.328253 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-x95rg" event={"ID":"4ca1cd05-e8fc-47b7-904b-a0a544f654b0","Type":"ContainerDied","Data":"d464b169c60197b9c240cfa168fe7b6ef35f19e0d369fca2260676bae19969cc"} Mar 17 09:14:38 crc kubenswrapper[4813]: I0317 09:14:38.328295 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-x95rg" Mar 17 09:14:38 crc kubenswrapper[4813]: I0317 09:14:38.328307 4813 scope.go:117] "RemoveContainer" containerID="988ee7f8b66efd6561f1e628375cfe7016af93e9579b928a9f95754744bb9ac7" Mar 17 09:14:38 crc kubenswrapper[4813]: I0317 09:14:38.330954 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5lhdz" event={"ID":"404fc67e-83be-4be6-bf6e-4279c0543316","Type":"ContainerStarted","Data":"9418ad0fa6a04609fe166c888da8c9e4a35a4a91b13726200a9a89bb54f661be"} Mar 17 09:14:38 crc kubenswrapper[4813]: I0317 09:14:38.343673 4813 scope.go:117] "RemoveContainer" containerID="ba6c548eb755d9564ab731194956b862e51e6e871050ad96f7f70efec7ed5611" Mar 17 09:14:38 crc kubenswrapper[4813]: I0317 09:14:38.362459 4813 scope.go:117] "RemoveContainer" containerID="41421cebc179c392b9464e16fdc9c15c1c1e7aef5e6ba069628a436d12d3ed22" Mar 17 09:14:38 crc kubenswrapper[4813]: I0317 09:14:38.371295 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5lhdz" podStartSLOduration=2.9106131729999998 podStartE2EDuration="47.371281472s" podCreationTimestamp="2026-03-17 09:13:51 +0000 UTC" firstStartedPulling="2026-03-17 09:13:53.618238127 +0000 UTC m=+255.719041626" lastFinishedPulling="2026-03-17 09:14:38.078906406 +0000 UTC m=+300.179709925" observedRunningTime="2026-03-17 09:14:38.368732156 +0000 UTC m=+300.469535655" watchObservedRunningTime="2026-03-17 09:14:38.371281472 +0000 UTC m=+300.472084971" Mar 17 09:14:38 crc kubenswrapper[4813]: I0317 09:14:38.381654 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-x95rg"] Mar 17 09:14:38 crc kubenswrapper[4813]: I0317 09:14:38.385088 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-x95rg"] Mar 17 09:14:38 crc kubenswrapper[4813]: I0317 09:14:38.746996 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ca1cd05-e8fc-47b7-904b-a0a544f654b0" path="/var/lib/kubelet/pods/4ca1cd05-e8fc-47b7-904b-a0a544f654b0/volumes" Mar 17 09:14:41 crc kubenswrapper[4813]: I0317 09:14:41.682831 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-lxrc7" Mar 17 09:14:41 crc kubenswrapper[4813]: I0317 09:14:41.683174 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-lxrc7" Mar 17 09:14:41 crc kubenswrapper[4813]: I0317 09:14:41.725152 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-lxrc7" Mar 17 09:14:42 crc kubenswrapper[4813]: I0317 09:14:42.066469 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5lhdz" Mar 17 09:14:42 crc kubenswrapper[4813]: I0317 09:14:42.066859 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5lhdz" Mar 17 09:14:42 crc kubenswrapper[4813]: I0317 09:14:42.120845 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5lhdz" Mar 17 09:14:42 crc kubenswrapper[4813]: I0317 09:14:42.410759 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-lxrc7" Mar 17 09:14:44 crc kubenswrapper[4813]: I0317 09:14:44.114029 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:14:44 crc kubenswrapper[4813]: I0317 09:14:44.114092 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:14:44 crc kubenswrapper[4813]: I0317 09:14:44.114157 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:14:44 crc kubenswrapper[4813]: I0317 09:14:44.115100 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09"} pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 09:14:44 crc kubenswrapper[4813]: I0317 09:14:44.115389 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" containerID="cri-o://1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09" gracePeriod=600 Mar 17 09:14:44 crc kubenswrapper[4813]: I0317 09:14:44.759160 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6gvsx" Mar 17 09:14:44 crc kubenswrapper[4813]: I0317 09:14:44.808991 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6gvsx" Mar 17 09:14:45 crc kubenswrapper[4813]: I0317 09:14:45.156692 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-bf7m2" Mar 17 09:14:45 crc kubenswrapper[4813]: I0317 09:14:45.209428 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-bf7m2" Mar 17 09:14:45 crc kubenswrapper[4813]: I0317 09:14:45.375233 4813 generic.go:334] "Generic (PLEG): container finished" podID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerID="1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09" exitCode=0 Mar 17 09:14:45 crc kubenswrapper[4813]: I0317 09:14:45.375336 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerDied","Data":"1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09"} Mar 17 09:14:45 crc kubenswrapper[4813]: I0317 09:14:45.375400 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerStarted","Data":"af70be0dbae35ba7879eb8d4732f829e536191eac3391f7a3a43fe520584a7bf"} Mar 17 09:14:47 crc kubenswrapper[4813]: I0317 09:14:47.573839 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bf7m2"] Mar 17 09:14:47 crc kubenswrapper[4813]: I0317 09:14:47.574392 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-bf7m2" podUID="e49eaa7a-38bf-4996-ad34-64959cbccda2" containerName="registry-server" containerID="cri-o://150c711cf5147379c0e341ae0163f834f4edcafe7ceea1557024f3432cfd2645" gracePeriod=2 Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.086854 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bf7m2" Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.269961 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e49eaa7a-38bf-4996-ad34-64959cbccda2-catalog-content\") pod \"e49eaa7a-38bf-4996-ad34-64959cbccda2\" (UID: \"e49eaa7a-38bf-4996-ad34-64959cbccda2\") " Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.270114 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e49eaa7a-38bf-4996-ad34-64959cbccda2-utilities\") pod \"e49eaa7a-38bf-4996-ad34-64959cbccda2\" (UID: \"e49eaa7a-38bf-4996-ad34-64959cbccda2\") " Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.270177 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blzbh\" (UniqueName: \"kubernetes.io/projected/e49eaa7a-38bf-4996-ad34-64959cbccda2-kube-api-access-blzbh\") pod \"e49eaa7a-38bf-4996-ad34-64959cbccda2\" (UID: \"e49eaa7a-38bf-4996-ad34-64959cbccda2\") " Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.272292 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e49eaa7a-38bf-4996-ad34-64959cbccda2-utilities" (OuterVolumeSpecName: "utilities") pod "e49eaa7a-38bf-4996-ad34-64959cbccda2" (UID: "e49eaa7a-38bf-4996-ad34-64959cbccda2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.276080 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e49eaa7a-38bf-4996-ad34-64959cbccda2-kube-api-access-blzbh" (OuterVolumeSpecName: "kube-api-access-blzbh") pod "e49eaa7a-38bf-4996-ad34-64959cbccda2" (UID: "e49eaa7a-38bf-4996-ad34-64959cbccda2"). InnerVolumeSpecName "kube-api-access-blzbh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.373066 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e49eaa7a-38bf-4996-ad34-64959cbccda2-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.373110 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blzbh\" (UniqueName: \"kubernetes.io/projected/e49eaa7a-38bf-4996-ad34-64959cbccda2-kube-api-access-blzbh\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.398899 4813 generic.go:334] "Generic (PLEG): container finished" podID="e49eaa7a-38bf-4996-ad34-64959cbccda2" containerID="150c711cf5147379c0e341ae0163f834f4edcafe7ceea1557024f3432cfd2645" exitCode=0 Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.398966 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bf7m2" event={"ID":"e49eaa7a-38bf-4996-ad34-64959cbccda2","Type":"ContainerDied","Data":"150c711cf5147379c0e341ae0163f834f4edcafe7ceea1557024f3432cfd2645"} Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.399026 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-bf7m2" event={"ID":"e49eaa7a-38bf-4996-ad34-64959cbccda2","Type":"ContainerDied","Data":"9fdffdd82d65d337053a38a5ed253a94699ac97963d750f597c483bca38d1043"} Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.399031 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-bf7m2" Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.399064 4813 scope.go:117] "RemoveContainer" containerID="150c711cf5147379c0e341ae0163f834f4edcafe7ceea1557024f3432cfd2645" Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.417934 4813 scope.go:117] "RemoveContainer" containerID="4395a333d2cec6f2d56a2627f6cf0133dc3089cf05296bf74d047fc950f1d147" Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.432464 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e49eaa7a-38bf-4996-ad34-64959cbccda2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e49eaa7a-38bf-4996-ad34-64959cbccda2" (UID: "e49eaa7a-38bf-4996-ad34-64959cbccda2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.441743 4813 scope.go:117] "RemoveContainer" containerID="27d9c820b252b46f72190dc09d97f814a6827d85e54fe13cb7ebc5a56f9b27f0" Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.459131 4813 scope.go:117] "RemoveContainer" containerID="150c711cf5147379c0e341ae0163f834f4edcafe7ceea1557024f3432cfd2645" Mar 17 09:14:48 crc kubenswrapper[4813]: E0317 09:14:48.459632 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"150c711cf5147379c0e341ae0163f834f4edcafe7ceea1557024f3432cfd2645\": container with ID starting with 150c711cf5147379c0e341ae0163f834f4edcafe7ceea1557024f3432cfd2645 not found: ID does not exist" containerID="150c711cf5147379c0e341ae0163f834f4edcafe7ceea1557024f3432cfd2645" Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.459670 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"150c711cf5147379c0e341ae0163f834f4edcafe7ceea1557024f3432cfd2645"} err="failed to get container status \"150c711cf5147379c0e341ae0163f834f4edcafe7ceea1557024f3432cfd2645\": rpc error: code = NotFound desc = could not find container \"150c711cf5147379c0e341ae0163f834f4edcafe7ceea1557024f3432cfd2645\": container with ID starting with 150c711cf5147379c0e341ae0163f834f4edcafe7ceea1557024f3432cfd2645 not found: ID does not exist" Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.459697 4813 scope.go:117] "RemoveContainer" containerID="4395a333d2cec6f2d56a2627f6cf0133dc3089cf05296bf74d047fc950f1d147" Mar 17 09:14:48 crc kubenswrapper[4813]: E0317 09:14:48.460148 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4395a333d2cec6f2d56a2627f6cf0133dc3089cf05296bf74d047fc950f1d147\": container with ID starting with 4395a333d2cec6f2d56a2627f6cf0133dc3089cf05296bf74d047fc950f1d147 not found: ID does not exist" containerID="4395a333d2cec6f2d56a2627f6cf0133dc3089cf05296bf74d047fc950f1d147" Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.460196 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4395a333d2cec6f2d56a2627f6cf0133dc3089cf05296bf74d047fc950f1d147"} err="failed to get container status \"4395a333d2cec6f2d56a2627f6cf0133dc3089cf05296bf74d047fc950f1d147\": rpc error: code = NotFound desc = could not find container \"4395a333d2cec6f2d56a2627f6cf0133dc3089cf05296bf74d047fc950f1d147\": container with ID starting with 4395a333d2cec6f2d56a2627f6cf0133dc3089cf05296bf74d047fc950f1d147 not found: ID does not exist" Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.460225 4813 scope.go:117] "RemoveContainer" containerID="27d9c820b252b46f72190dc09d97f814a6827d85e54fe13cb7ebc5a56f9b27f0" Mar 17 09:14:48 crc kubenswrapper[4813]: E0317 09:14:48.460671 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27d9c820b252b46f72190dc09d97f814a6827d85e54fe13cb7ebc5a56f9b27f0\": container with ID starting with 27d9c820b252b46f72190dc09d97f814a6827d85e54fe13cb7ebc5a56f9b27f0 not found: ID does not exist" containerID="27d9c820b252b46f72190dc09d97f814a6827d85e54fe13cb7ebc5a56f9b27f0" Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.460741 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27d9c820b252b46f72190dc09d97f814a6827d85e54fe13cb7ebc5a56f9b27f0"} err="failed to get container status \"27d9c820b252b46f72190dc09d97f814a6827d85e54fe13cb7ebc5a56f9b27f0\": rpc error: code = NotFound desc = could not find container \"27d9c820b252b46f72190dc09d97f814a6827d85e54fe13cb7ebc5a56f9b27f0\": container with ID starting with 27d9c820b252b46f72190dc09d97f814a6827d85e54fe13cb7ebc5a56f9b27f0 not found: ID does not exist" Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.475119 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e49eaa7a-38bf-4996-ad34-64959cbccda2-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.756490 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-bf7m2"] Mar 17 09:14:48 crc kubenswrapper[4813]: I0317 09:14:48.760295 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-bf7m2"] Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.222456 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd"] Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.223007 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" podUID="788a513c-9124-4152-908c-df0c059b6270" containerName="controller-manager" containerID="cri-o://346f1361a81b5452be41bbe4ac74ab25262853f7c34d73b5f1b23d272f77ade6" gracePeriod=30 Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.245660 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz"] Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.246106 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" podUID="48a312b3-3ba5-40ab-bbc9-b4b21903f978" containerName="route-controller-manager" containerID="cri-o://aa7b6baf3551238cdeea1f20748a58f0b60f5a023e3f01815149a05c4fe21c1d" gracePeriod=30 Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.406291 4813 generic.go:334] "Generic (PLEG): container finished" podID="788a513c-9124-4152-908c-df0c059b6270" containerID="346f1361a81b5452be41bbe4ac74ab25262853f7c34d73b5f1b23d272f77ade6" exitCode=0 Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.406381 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" event={"ID":"788a513c-9124-4152-908c-df0c059b6270","Type":"ContainerDied","Data":"346f1361a81b5452be41bbe4ac74ab25262853f7c34d73b5f1b23d272f77ade6"} Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.408768 4813 generic.go:334] "Generic (PLEG): container finished" podID="48a312b3-3ba5-40ab-bbc9-b4b21903f978" containerID="aa7b6baf3551238cdeea1f20748a58f0b60f5a023e3f01815149a05c4fe21c1d" exitCode=0 Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.408858 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" event={"ID":"48a312b3-3ba5-40ab-bbc9-b4b21903f978","Type":"ContainerDied","Data":"aa7b6baf3551238cdeea1f20748a58f0b60f5a023e3f01815149a05c4fe21c1d"} Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.755641 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.793709 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/48a312b3-3ba5-40ab-bbc9-b4b21903f978-client-ca\") pod \"48a312b3-3ba5-40ab-bbc9-b4b21903f978\" (UID: \"48a312b3-3ba5-40ab-bbc9-b4b21903f978\") " Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.793889 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48a312b3-3ba5-40ab-bbc9-b4b21903f978-serving-cert\") pod \"48a312b3-3ba5-40ab-bbc9-b4b21903f978\" (UID: \"48a312b3-3ba5-40ab-bbc9-b4b21903f978\") " Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.793943 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48a312b3-3ba5-40ab-bbc9-b4b21903f978-config\") pod \"48a312b3-3ba5-40ab-bbc9-b4b21903f978\" (UID: \"48a312b3-3ba5-40ab-bbc9-b4b21903f978\") " Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.793998 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28qpw\" (UniqueName: \"kubernetes.io/projected/48a312b3-3ba5-40ab-bbc9-b4b21903f978-kube-api-access-28qpw\") pod \"48a312b3-3ba5-40ab-bbc9-b4b21903f978\" (UID: \"48a312b3-3ba5-40ab-bbc9-b4b21903f978\") " Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.794508 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48a312b3-3ba5-40ab-bbc9-b4b21903f978-client-ca" (OuterVolumeSpecName: "client-ca") pod "48a312b3-3ba5-40ab-bbc9-b4b21903f978" (UID: "48a312b3-3ba5-40ab-bbc9-b4b21903f978"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.794934 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48a312b3-3ba5-40ab-bbc9-b4b21903f978-config" (OuterVolumeSpecName: "config") pod "48a312b3-3ba5-40ab-bbc9-b4b21903f978" (UID: "48a312b3-3ba5-40ab-bbc9-b4b21903f978"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.801846 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48a312b3-3ba5-40ab-bbc9-b4b21903f978-kube-api-access-28qpw" (OuterVolumeSpecName: "kube-api-access-28qpw") pod "48a312b3-3ba5-40ab-bbc9-b4b21903f978" (UID: "48a312b3-3ba5-40ab-bbc9-b4b21903f978"). InnerVolumeSpecName "kube-api-access-28qpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.802014 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48a312b3-3ba5-40ab-bbc9-b4b21903f978-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "48a312b3-3ba5-40ab-bbc9-b4b21903f978" (UID: "48a312b3-3ba5-40ab-bbc9-b4b21903f978"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.851243 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.894856 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/788a513c-9124-4152-908c-df0c059b6270-client-ca\") pod \"788a513c-9124-4152-908c-df0c059b6270\" (UID: \"788a513c-9124-4152-908c-df0c059b6270\") " Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.894931 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/788a513c-9124-4152-908c-df0c059b6270-proxy-ca-bundles\") pod \"788a513c-9124-4152-908c-df0c059b6270\" (UID: \"788a513c-9124-4152-908c-df0c059b6270\") " Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.895010 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/788a513c-9124-4152-908c-df0c059b6270-config\") pod \"788a513c-9124-4152-908c-df0c059b6270\" (UID: \"788a513c-9124-4152-908c-df0c059b6270\") " Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.895052 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/788a513c-9124-4152-908c-df0c059b6270-serving-cert\") pod \"788a513c-9124-4152-908c-df0c059b6270\" (UID: \"788a513c-9124-4152-908c-df0c059b6270\") " Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.895160 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvqkb\" (UniqueName: \"kubernetes.io/projected/788a513c-9124-4152-908c-df0c059b6270-kube-api-access-vvqkb\") pod \"788a513c-9124-4152-908c-df0c059b6270\" (UID: \"788a513c-9124-4152-908c-df0c059b6270\") " Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.895487 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/48a312b3-3ba5-40ab-bbc9-b4b21903f978-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.895512 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48a312b3-3ba5-40ab-bbc9-b4b21903f978-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.895531 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28qpw\" (UniqueName: \"kubernetes.io/projected/48a312b3-3ba5-40ab-bbc9-b4b21903f978-kube-api-access-28qpw\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.895551 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/48a312b3-3ba5-40ab-bbc9-b4b21903f978-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.896356 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/788a513c-9124-4152-908c-df0c059b6270-client-ca" (OuterVolumeSpecName: "client-ca") pod "788a513c-9124-4152-908c-df0c059b6270" (UID: "788a513c-9124-4152-908c-df0c059b6270"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.896452 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/788a513c-9124-4152-908c-df0c059b6270-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "788a513c-9124-4152-908c-df0c059b6270" (UID: "788a513c-9124-4152-908c-df0c059b6270"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.896949 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/788a513c-9124-4152-908c-df0c059b6270-config" (OuterVolumeSpecName: "config") pod "788a513c-9124-4152-908c-df0c059b6270" (UID: "788a513c-9124-4152-908c-df0c059b6270"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.898967 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/788a513c-9124-4152-908c-df0c059b6270-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "788a513c-9124-4152-908c-df0c059b6270" (UID: "788a513c-9124-4152-908c-df0c059b6270"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.899565 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/788a513c-9124-4152-908c-df0c059b6270-kube-api-access-vvqkb" (OuterVolumeSpecName: "kube-api-access-vvqkb") pod "788a513c-9124-4152-908c-df0c059b6270" (UID: "788a513c-9124-4152-908c-df0c059b6270"). InnerVolumeSpecName "kube-api-access-vvqkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.997253 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/788a513c-9124-4152-908c-df0c059b6270-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.997288 4813 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/788a513c-9124-4152-908c-df0c059b6270-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.997304 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/788a513c-9124-4152-908c-df0c059b6270-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.997315 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/788a513c-9124-4152-908c-df0c059b6270-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:49 crc kubenswrapper[4813]: I0317 09:14:49.997326 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvqkb\" (UniqueName: \"kubernetes.io/projected/788a513c-9124-4152-908c-df0c059b6270-kube-api-access-vvqkb\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.418711 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" event={"ID":"788a513c-9124-4152-908c-df0c059b6270","Type":"ContainerDied","Data":"80da5c48cabb45575212074b1db4381bb00fd081cf4ef39fb6c96c3fa2b583ab"} Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.418748 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.418765 4813 scope.go:117] "RemoveContainer" containerID="346f1361a81b5452be41bbe4ac74ab25262853f7c34d73b5f1b23d272f77ade6" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.420483 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" event={"ID":"48a312b3-3ba5-40ab-bbc9-b4b21903f978","Type":"ContainerDied","Data":"2363b0c6a8f8f147e80ec520f30457a83fcd62b102dc4e4dd084b7f7d91ec6e1"} Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.420534 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.446634 4813 scope.go:117] "RemoveContainer" containerID="aa7b6baf3551238cdeea1f20748a58f0b60f5a023e3f01815149a05c4fe21c1d" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.457354 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz"] Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.465674 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-fd576fb67-mvgcz"] Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.478183 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd"] Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.485015 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7d655f7fcd-rrwmd"] Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.739317 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48a312b3-3ba5-40ab-bbc9-b4b21903f978" path="/var/lib/kubelet/pods/48a312b3-3ba5-40ab-bbc9-b4b21903f978/volumes" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.740097 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="788a513c-9124-4152-908c-df0c059b6270" path="/var/lib/kubelet/pods/788a513c-9124-4152-908c-df0c059b6270/volumes" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.740722 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e49eaa7a-38bf-4996-ad34-64959cbccda2" path="/var/lib/kubelet/pods/e49eaa7a-38bf-4996-ad34-64959cbccda2/volumes" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.950728 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7549dfdb87-dkkng"] Mar 17 09:14:50 crc kubenswrapper[4813]: E0317 09:14:50.951405 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ca1cd05-e8fc-47b7-904b-a0a544f654b0" containerName="extract-content" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.951426 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ca1cd05-e8fc-47b7-904b-a0a544f654b0" containerName="extract-content" Mar 17 09:14:50 crc kubenswrapper[4813]: E0317 09:14:50.951444 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e49eaa7a-38bf-4996-ad34-64959cbccda2" containerName="registry-server" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.951454 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e49eaa7a-38bf-4996-ad34-64959cbccda2" containerName="registry-server" Mar 17 09:14:50 crc kubenswrapper[4813]: E0317 09:14:50.951466 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e49eaa7a-38bf-4996-ad34-64959cbccda2" containerName="extract-utilities" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.951477 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e49eaa7a-38bf-4996-ad34-64959cbccda2" containerName="extract-utilities" Mar 17 09:14:50 crc kubenswrapper[4813]: E0317 09:14:50.951491 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ca1cd05-e8fc-47b7-904b-a0a544f654b0" containerName="registry-server" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.951501 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ca1cd05-e8fc-47b7-904b-a0a544f654b0" containerName="registry-server" Mar 17 09:14:50 crc kubenswrapper[4813]: E0317 09:14:50.951514 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e49eaa7a-38bf-4996-ad34-64959cbccda2" containerName="extract-content" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.951524 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e49eaa7a-38bf-4996-ad34-64959cbccda2" containerName="extract-content" Mar 17 09:14:50 crc kubenswrapper[4813]: E0317 09:14:50.951539 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7d1b948-0f9a-4ba3-9f58-88590e13ba63" containerName="extract-utilities" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.951550 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7d1b948-0f9a-4ba3-9f58-88590e13ba63" containerName="extract-utilities" Mar 17 09:14:50 crc kubenswrapper[4813]: E0317 09:14:50.951565 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="788a513c-9124-4152-908c-df0c059b6270" containerName="controller-manager" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.951573 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="788a513c-9124-4152-908c-df0c059b6270" containerName="controller-manager" Mar 17 09:14:50 crc kubenswrapper[4813]: E0317 09:14:50.951587 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ca1cd05-e8fc-47b7-904b-a0a544f654b0" containerName="extract-utilities" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.951597 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ca1cd05-e8fc-47b7-904b-a0a544f654b0" containerName="extract-utilities" Mar 17 09:14:50 crc kubenswrapper[4813]: E0317 09:14:50.951631 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7d1b948-0f9a-4ba3-9f58-88590e13ba63" containerName="extract-content" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.951641 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7d1b948-0f9a-4ba3-9f58-88590e13ba63" containerName="extract-content" Mar 17 09:14:50 crc kubenswrapper[4813]: E0317 09:14:50.951653 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7d1b948-0f9a-4ba3-9f58-88590e13ba63" containerName="registry-server" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.951664 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7d1b948-0f9a-4ba3-9f58-88590e13ba63" containerName="registry-server" Mar 17 09:14:50 crc kubenswrapper[4813]: E0317 09:14:50.951675 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48a312b3-3ba5-40ab-bbc9-b4b21903f978" containerName="route-controller-manager" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.951683 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="48a312b3-3ba5-40ab-bbc9-b4b21903f978" containerName="route-controller-manager" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.951816 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ca1cd05-e8fc-47b7-904b-a0a544f654b0" containerName="registry-server" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.951833 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7d1b948-0f9a-4ba3-9f58-88590e13ba63" containerName="registry-server" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.951849 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="788a513c-9124-4152-908c-df0c059b6270" containerName="controller-manager" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.951869 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="48a312b3-3ba5-40ab-bbc9-b4b21903f978" containerName="route-controller-manager" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.951886 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e49eaa7a-38bf-4996-ad34-64959cbccda2" containerName="registry-server" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.957152 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.958347 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8"] Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.960407 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" Mar 17 09:14:50 crc kubenswrapper[4813]: I0317 09:14:50.967755 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8"] Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.004408 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.005182 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.006251 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.006806 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.006937 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.006962 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.007367 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.007498 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.007188 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.007712 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.007804 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.007964 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.008296 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7549dfdb87-dkkng"] Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.017477 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.017557 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3261033c-f194-459b-952a-79391f6b3ea9-client-ca\") pod \"route-controller-manager-7767bbd8d-lw8t8\" (UID: \"3261033c-f194-459b-952a-79391f6b3ea9\") " pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.018455 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3261033c-f194-459b-952a-79391f6b3ea9-config\") pod \"route-controller-manager-7767bbd8d-lw8t8\" (UID: \"3261033c-f194-459b-952a-79391f6b3ea9\") " pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.018926 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7207af26-0b14-4ce2-b178-2daad0226706-proxy-ca-bundles\") pod \"controller-manager-7549dfdb87-dkkng\" (UID: \"7207af26-0b14-4ce2-b178-2daad0226706\") " pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.019156 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7207af26-0b14-4ce2-b178-2daad0226706-serving-cert\") pod \"controller-manager-7549dfdb87-dkkng\" (UID: \"7207af26-0b14-4ce2-b178-2daad0226706\") " pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.019648 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5bgj\" (UniqueName: \"kubernetes.io/projected/7207af26-0b14-4ce2-b178-2daad0226706-kube-api-access-g5bgj\") pod \"controller-manager-7549dfdb87-dkkng\" (UID: \"7207af26-0b14-4ce2-b178-2daad0226706\") " pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.019871 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w87z\" (UniqueName: \"kubernetes.io/projected/3261033c-f194-459b-952a-79391f6b3ea9-kube-api-access-6w87z\") pod \"route-controller-manager-7767bbd8d-lw8t8\" (UID: \"3261033c-f194-459b-952a-79391f6b3ea9\") " pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.020068 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3261033c-f194-459b-952a-79391f6b3ea9-serving-cert\") pod \"route-controller-manager-7767bbd8d-lw8t8\" (UID: \"3261033c-f194-459b-952a-79391f6b3ea9\") " pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.020262 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7207af26-0b14-4ce2-b178-2daad0226706-config\") pod \"controller-manager-7549dfdb87-dkkng\" (UID: \"7207af26-0b14-4ce2-b178-2daad0226706\") " pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.020429 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7207af26-0b14-4ce2-b178-2daad0226706-client-ca\") pod \"controller-manager-7549dfdb87-dkkng\" (UID: \"7207af26-0b14-4ce2-b178-2daad0226706\") " pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.122159 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7207af26-0b14-4ce2-b178-2daad0226706-serving-cert\") pod \"controller-manager-7549dfdb87-dkkng\" (UID: \"7207af26-0b14-4ce2-b178-2daad0226706\") " pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.122208 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5bgj\" (UniqueName: \"kubernetes.io/projected/7207af26-0b14-4ce2-b178-2daad0226706-kube-api-access-g5bgj\") pod \"controller-manager-7549dfdb87-dkkng\" (UID: \"7207af26-0b14-4ce2-b178-2daad0226706\") " pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.122240 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w87z\" (UniqueName: \"kubernetes.io/projected/3261033c-f194-459b-952a-79391f6b3ea9-kube-api-access-6w87z\") pod \"route-controller-manager-7767bbd8d-lw8t8\" (UID: \"3261033c-f194-459b-952a-79391f6b3ea9\") " pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.122271 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3261033c-f194-459b-952a-79391f6b3ea9-serving-cert\") pod \"route-controller-manager-7767bbd8d-lw8t8\" (UID: \"3261033c-f194-459b-952a-79391f6b3ea9\") " pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.122298 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7207af26-0b14-4ce2-b178-2daad0226706-config\") pod \"controller-manager-7549dfdb87-dkkng\" (UID: \"7207af26-0b14-4ce2-b178-2daad0226706\") " pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.122317 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7207af26-0b14-4ce2-b178-2daad0226706-client-ca\") pod \"controller-manager-7549dfdb87-dkkng\" (UID: \"7207af26-0b14-4ce2-b178-2daad0226706\") " pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.122354 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3261033c-f194-459b-952a-79391f6b3ea9-client-ca\") pod \"route-controller-manager-7767bbd8d-lw8t8\" (UID: \"3261033c-f194-459b-952a-79391f6b3ea9\") " pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.122376 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3261033c-f194-459b-952a-79391f6b3ea9-config\") pod \"route-controller-manager-7767bbd8d-lw8t8\" (UID: \"3261033c-f194-459b-952a-79391f6b3ea9\") " pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.122395 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7207af26-0b14-4ce2-b178-2daad0226706-proxy-ca-bundles\") pod \"controller-manager-7549dfdb87-dkkng\" (UID: \"7207af26-0b14-4ce2-b178-2daad0226706\") " pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.123514 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7207af26-0b14-4ce2-b178-2daad0226706-proxy-ca-bundles\") pod \"controller-manager-7549dfdb87-dkkng\" (UID: \"7207af26-0b14-4ce2-b178-2daad0226706\") " pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.123753 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3261033c-f194-459b-952a-79391f6b3ea9-client-ca\") pod \"route-controller-manager-7767bbd8d-lw8t8\" (UID: \"3261033c-f194-459b-952a-79391f6b3ea9\") " pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.123841 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3261033c-f194-459b-952a-79391f6b3ea9-config\") pod \"route-controller-manager-7767bbd8d-lw8t8\" (UID: \"3261033c-f194-459b-952a-79391f6b3ea9\") " pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.124092 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7207af26-0b14-4ce2-b178-2daad0226706-config\") pod \"controller-manager-7549dfdb87-dkkng\" (UID: \"7207af26-0b14-4ce2-b178-2daad0226706\") " pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.124469 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7207af26-0b14-4ce2-b178-2daad0226706-client-ca\") pod \"controller-manager-7549dfdb87-dkkng\" (UID: \"7207af26-0b14-4ce2-b178-2daad0226706\") " pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.128237 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7207af26-0b14-4ce2-b178-2daad0226706-serving-cert\") pod \"controller-manager-7549dfdb87-dkkng\" (UID: \"7207af26-0b14-4ce2-b178-2daad0226706\") " pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.132702 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3261033c-f194-459b-952a-79391f6b3ea9-serving-cert\") pod \"route-controller-manager-7767bbd8d-lw8t8\" (UID: \"3261033c-f194-459b-952a-79391f6b3ea9\") " pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.155559 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5bgj\" (UniqueName: \"kubernetes.io/projected/7207af26-0b14-4ce2-b178-2daad0226706-kube-api-access-g5bgj\") pod \"controller-manager-7549dfdb87-dkkng\" (UID: \"7207af26-0b14-4ce2-b178-2daad0226706\") " pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.163743 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w87z\" (UniqueName: \"kubernetes.io/projected/3261033c-f194-459b-952a-79391f6b3ea9-kube-api-access-6w87z\") pod \"route-controller-manager-7767bbd8d-lw8t8\" (UID: \"3261033c-f194-459b-952a-79391f6b3ea9\") " pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.315308 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.322964 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.575431 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7549dfdb87-dkkng"] Mar 17 09:14:51 crc kubenswrapper[4813]: W0317 09:14:51.588414 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7207af26_0b14_4ce2_b178_2daad0226706.slice/crio-cc2ab0d3a32b145182a81ceabe5cfba9debd88314f18399b00f8e068e5b28bee WatchSource:0}: Error finding container cc2ab0d3a32b145182a81ceabe5cfba9debd88314f18399b00f8e068e5b28bee: Status 404 returned error can't find the container with id cc2ab0d3a32b145182a81ceabe5cfba9debd88314f18399b00f8e068e5b28bee Mar 17 09:14:51 crc kubenswrapper[4813]: I0317 09:14:51.843158 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8"] Mar 17 09:14:51 crc kubenswrapper[4813]: W0317 09:14:51.849587 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3261033c_f194_459b_952a_79391f6b3ea9.slice/crio-3fb614813cba56cbb39aeb44d7a979eb10a9c9c30087a17b2ab0ad70d1c039fd WatchSource:0}: Error finding container 3fb614813cba56cbb39aeb44d7a979eb10a9c9c30087a17b2ab0ad70d1c039fd: Status 404 returned error can't find the container with id 3fb614813cba56cbb39aeb44d7a979eb10a9c9c30087a17b2ab0ad70d1c039fd Mar 17 09:14:52 crc kubenswrapper[4813]: I0317 09:14:52.109133 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5lhdz" Mar 17 09:14:52 crc kubenswrapper[4813]: I0317 09:14:52.439006 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" event={"ID":"7207af26-0b14-4ce2-b178-2daad0226706","Type":"ContainerStarted","Data":"6a9a555ebe60ef6ea5e88a305c1e1dd5899eb32ec794cb61aa4c77b8a8f77da3"} Mar 17 09:14:52 crc kubenswrapper[4813]: I0317 09:14:52.439079 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" event={"ID":"7207af26-0b14-4ce2-b178-2daad0226706","Type":"ContainerStarted","Data":"cc2ab0d3a32b145182a81ceabe5cfba9debd88314f18399b00f8e068e5b28bee"} Mar 17 09:14:52 crc kubenswrapper[4813]: I0317 09:14:52.439117 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" Mar 17 09:14:52 crc kubenswrapper[4813]: I0317 09:14:52.440977 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" event={"ID":"3261033c-f194-459b-952a-79391f6b3ea9","Type":"ContainerStarted","Data":"c5c9eb13ac2c4ae5a424a8105d447b12e53d0dcfc5d76e9abba0ff72c354c798"} Mar 17 09:14:52 crc kubenswrapper[4813]: I0317 09:14:52.441058 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" event={"ID":"3261033c-f194-459b-952a-79391f6b3ea9","Type":"ContainerStarted","Data":"3fb614813cba56cbb39aeb44d7a979eb10a9c9c30087a17b2ab0ad70d1c039fd"} Mar 17 09:14:52 crc kubenswrapper[4813]: I0317 09:14:52.441222 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" Mar 17 09:14:52 crc kubenswrapper[4813]: I0317 09:14:52.449523 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" Mar 17 09:14:52 crc kubenswrapper[4813]: I0317 09:14:52.463805 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" podStartSLOduration=3.46378091 podStartE2EDuration="3.46378091s" podCreationTimestamp="2026-03-17 09:14:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:14:52.461788181 +0000 UTC m=+314.562591720" watchObservedRunningTime="2026-03-17 09:14:52.46378091 +0000 UTC m=+314.564584429" Mar 17 09:14:52 crc kubenswrapper[4813]: I0317 09:14:52.497290 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" podStartSLOduration=3.49727345 podStartE2EDuration="3.49727345s" podCreationTimestamp="2026-03-17 09:14:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:14:52.496414041 +0000 UTC m=+314.597217540" watchObservedRunningTime="2026-03-17 09:14:52.49727345 +0000 UTC m=+314.598076949" Mar 17 09:14:52 crc kubenswrapper[4813]: I0317 09:14:52.652290 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" Mar 17 09:14:53 crc kubenswrapper[4813]: I0317 09:14:53.575927 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5lhdz"] Mar 17 09:14:53 crc kubenswrapper[4813]: I0317 09:14:53.576740 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5lhdz" podUID="404fc67e-83be-4be6-bf6e-4279c0543316" containerName="registry-server" containerID="cri-o://9418ad0fa6a04609fe166c888da8c9e4a35a4a91b13726200a9a89bb54f661be" gracePeriod=2 Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.013381 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5lhdz" Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.060419 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7vwj\" (UniqueName: \"kubernetes.io/projected/404fc67e-83be-4be6-bf6e-4279c0543316-kube-api-access-q7vwj\") pod \"404fc67e-83be-4be6-bf6e-4279c0543316\" (UID: \"404fc67e-83be-4be6-bf6e-4279c0543316\") " Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.060486 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/404fc67e-83be-4be6-bf6e-4279c0543316-utilities\") pod \"404fc67e-83be-4be6-bf6e-4279c0543316\" (UID: \"404fc67e-83be-4be6-bf6e-4279c0543316\") " Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.060527 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/404fc67e-83be-4be6-bf6e-4279c0543316-catalog-content\") pod \"404fc67e-83be-4be6-bf6e-4279c0543316\" (UID: \"404fc67e-83be-4be6-bf6e-4279c0543316\") " Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.061499 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/404fc67e-83be-4be6-bf6e-4279c0543316-utilities" (OuterVolumeSpecName: "utilities") pod "404fc67e-83be-4be6-bf6e-4279c0543316" (UID: "404fc67e-83be-4be6-bf6e-4279c0543316"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.067970 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/404fc67e-83be-4be6-bf6e-4279c0543316-kube-api-access-q7vwj" (OuterVolumeSpecName: "kube-api-access-q7vwj") pod "404fc67e-83be-4be6-bf6e-4279c0543316" (UID: "404fc67e-83be-4be6-bf6e-4279c0543316"). InnerVolumeSpecName "kube-api-access-q7vwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.118569 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/404fc67e-83be-4be6-bf6e-4279c0543316-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "404fc67e-83be-4be6-bf6e-4279c0543316" (UID: "404fc67e-83be-4be6-bf6e-4279c0543316"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.162272 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7vwj\" (UniqueName: \"kubernetes.io/projected/404fc67e-83be-4be6-bf6e-4279c0543316-kube-api-access-q7vwj\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.162313 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/404fc67e-83be-4be6-bf6e-4279c0543316-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.162322 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/404fc67e-83be-4be6-bf6e-4279c0543316-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.455948 4813 generic.go:334] "Generic (PLEG): container finished" podID="404fc67e-83be-4be6-bf6e-4279c0543316" containerID="9418ad0fa6a04609fe166c888da8c9e4a35a4a91b13726200a9a89bb54f661be" exitCode=0 Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.456009 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5lhdz" event={"ID":"404fc67e-83be-4be6-bf6e-4279c0543316","Type":"ContainerDied","Data":"9418ad0fa6a04609fe166c888da8c9e4a35a4a91b13726200a9a89bb54f661be"} Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.456066 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5lhdz" Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.456095 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5lhdz" event={"ID":"404fc67e-83be-4be6-bf6e-4279c0543316","Type":"ContainerDied","Data":"b60bb60cfafc59976f24900693c81a01dc2dbf5000f8873a6a74143b1ac0ad0b"} Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.456122 4813 scope.go:117] "RemoveContainer" containerID="9418ad0fa6a04609fe166c888da8c9e4a35a4a91b13726200a9a89bb54f661be" Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.481733 4813 scope.go:117] "RemoveContainer" containerID="2097079cc5e14a78c96d867fbb7b8d994d7f551495e284af871ea520ff284874" Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.499407 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5lhdz"] Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.502558 4813 scope.go:117] "RemoveContainer" containerID="906a9701eab35212e064aab915a49fbd3f1bf2006d8079b721076b375f00ff18" Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.507776 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5lhdz"] Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.536318 4813 scope.go:117] "RemoveContainer" containerID="9418ad0fa6a04609fe166c888da8c9e4a35a4a91b13726200a9a89bb54f661be" Mar 17 09:14:54 crc kubenswrapper[4813]: E0317 09:14:54.536853 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9418ad0fa6a04609fe166c888da8c9e4a35a4a91b13726200a9a89bb54f661be\": container with ID starting with 9418ad0fa6a04609fe166c888da8c9e4a35a4a91b13726200a9a89bb54f661be not found: ID does not exist" containerID="9418ad0fa6a04609fe166c888da8c9e4a35a4a91b13726200a9a89bb54f661be" Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.536921 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9418ad0fa6a04609fe166c888da8c9e4a35a4a91b13726200a9a89bb54f661be"} err="failed to get container status \"9418ad0fa6a04609fe166c888da8c9e4a35a4a91b13726200a9a89bb54f661be\": rpc error: code = NotFound desc = could not find container \"9418ad0fa6a04609fe166c888da8c9e4a35a4a91b13726200a9a89bb54f661be\": container with ID starting with 9418ad0fa6a04609fe166c888da8c9e4a35a4a91b13726200a9a89bb54f661be not found: ID does not exist" Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.536963 4813 scope.go:117] "RemoveContainer" containerID="2097079cc5e14a78c96d867fbb7b8d994d7f551495e284af871ea520ff284874" Mar 17 09:14:54 crc kubenswrapper[4813]: E0317 09:14:54.537693 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2097079cc5e14a78c96d867fbb7b8d994d7f551495e284af871ea520ff284874\": container with ID starting with 2097079cc5e14a78c96d867fbb7b8d994d7f551495e284af871ea520ff284874 not found: ID does not exist" containerID="2097079cc5e14a78c96d867fbb7b8d994d7f551495e284af871ea520ff284874" Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.537727 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2097079cc5e14a78c96d867fbb7b8d994d7f551495e284af871ea520ff284874"} err="failed to get container status \"2097079cc5e14a78c96d867fbb7b8d994d7f551495e284af871ea520ff284874\": rpc error: code = NotFound desc = could not find container \"2097079cc5e14a78c96d867fbb7b8d994d7f551495e284af871ea520ff284874\": container with ID starting with 2097079cc5e14a78c96d867fbb7b8d994d7f551495e284af871ea520ff284874 not found: ID does not exist" Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.537748 4813 scope.go:117] "RemoveContainer" containerID="906a9701eab35212e064aab915a49fbd3f1bf2006d8079b721076b375f00ff18" Mar 17 09:14:54 crc kubenswrapper[4813]: E0317 09:14:54.538098 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"906a9701eab35212e064aab915a49fbd3f1bf2006d8079b721076b375f00ff18\": container with ID starting with 906a9701eab35212e064aab915a49fbd3f1bf2006d8079b721076b375f00ff18 not found: ID does not exist" containerID="906a9701eab35212e064aab915a49fbd3f1bf2006d8079b721076b375f00ff18" Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.538162 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"906a9701eab35212e064aab915a49fbd3f1bf2006d8079b721076b375f00ff18"} err="failed to get container status \"906a9701eab35212e064aab915a49fbd3f1bf2006d8079b721076b375f00ff18\": rpc error: code = NotFound desc = could not find container \"906a9701eab35212e064aab915a49fbd3f1bf2006d8079b721076b375f00ff18\": container with ID starting with 906a9701eab35212e064aab915a49fbd3f1bf2006d8079b721076b375f00ff18 not found: ID does not exist" Mar 17 09:14:54 crc kubenswrapper[4813]: I0317 09:14:54.738033 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="404fc67e-83be-4be6-bf6e-4279c0543316" path="/var/lib/kubelet/pods/404fc67e-83be-4be6-bf6e-4279c0543316/volumes" Mar 17 09:14:57 crc kubenswrapper[4813]: I0317 09:14:57.916294 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" podUID="5e837d92-c48e-4b0f-a55a-40bdaea5d9f0" containerName="oauth-openshift" containerID="cri-o://a9484cecf8ed9d637c1ce56409a51a79ed0919e00abef61a98204d77038b3650" gracePeriod=15 Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.437051 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.484624 4813 generic.go:334] "Generic (PLEG): container finished" podID="5e837d92-c48e-4b0f-a55a-40bdaea5d9f0" containerID="a9484cecf8ed9d637c1ce56409a51a79ed0919e00abef61a98204d77038b3650" exitCode=0 Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.484687 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" event={"ID":"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0","Type":"ContainerDied","Data":"a9484cecf8ed9d637c1ce56409a51a79ed0919e00abef61a98204d77038b3650"} Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.484724 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" event={"ID":"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0","Type":"ContainerDied","Data":"0a9009f31556196a712dbea20111012eb1c927a32e77a31daaae2fc5e415da10"} Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.484753 4813 scope.go:117] "RemoveContainer" containerID="a9484cecf8ed9d637c1ce56409a51a79ed0919e00abef61a98204d77038b3650" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.484907 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-5kp7w" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.509690 4813 scope.go:117] "RemoveContainer" containerID="a9484cecf8ed9d637c1ce56409a51a79ed0919e00abef61a98204d77038b3650" Mar 17 09:14:58 crc kubenswrapper[4813]: E0317 09:14:58.510157 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9484cecf8ed9d637c1ce56409a51a79ed0919e00abef61a98204d77038b3650\": container with ID starting with a9484cecf8ed9d637c1ce56409a51a79ed0919e00abef61a98204d77038b3650 not found: ID does not exist" containerID="a9484cecf8ed9d637c1ce56409a51a79ed0919e00abef61a98204d77038b3650" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.510204 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9484cecf8ed9d637c1ce56409a51a79ed0919e00abef61a98204d77038b3650"} err="failed to get container status \"a9484cecf8ed9d637c1ce56409a51a79ed0919e00abef61a98204d77038b3650\": rpc error: code = NotFound desc = could not find container \"a9484cecf8ed9d637c1ce56409a51a79ed0919e00abef61a98204d77038b3650\": container with ID starting with a9484cecf8ed9d637c1ce56409a51a79ed0919e00abef61a98204d77038b3650 not found: ID does not exist" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.528843 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-trusted-ca-bundle\") pod \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.528888 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-template-login\") pod \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.528930 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-router-certs\") pod \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.528962 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cw28k\" (UniqueName: \"kubernetes.io/projected/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-kube-api-access-cw28k\") pod \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.529008 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-service-ca\") pod \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.529042 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-template-error\") pod \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.529069 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-audit-dir\") pod \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.529094 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-ocp-branding-template\") pod \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.529126 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-cliconfig\") pod \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.529146 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-serving-cert\") pod \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.529185 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-template-provider-selection\") pod \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.529233 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-audit-policies\") pod \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.529259 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-session\") pod \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.529287 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-idp-0-file-data\") pod \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\" (UID: \"5e837d92-c48e-4b0f-a55a-40bdaea5d9f0\") " Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.529922 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0" (UID: "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.529939 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0" (UID: "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.530105 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0" (UID: "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.530169 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0" (UID: "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.530776 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0" (UID: "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.536325 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0" (UID: "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.536332 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-kube-api-access-cw28k" (OuterVolumeSpecName: "kube-api-access-cw28k") pod "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0" (UID: "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0"). InnerVolumeSpecName "kube-api-access-cw28k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.536576 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0" (UID: "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.536825 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0" (UID: "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.537012 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0" (UID: "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.537413 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0" (UID: "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.542997 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0" (UID: "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.543121 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0" (UID: "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.543254 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0" (UID: "5e837d92-c48e-4b0f-a55a-40bdaea5d9f0"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.630806 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.630848 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cw28k\" (UniqueName: \"kubernetes.io/projected/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-kube-api-access-cw28k\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.630858 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.630869 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.630880 4813 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.630890 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.630900 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.630908 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.630920 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.630930 4813 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.630938 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.630947 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.630955 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.630963 4813 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.821092 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5kp7w"] Mar 17 09:14:58 crc kubenswrapper[4813]: I0317 09:14:58.827853 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-5kp7w"] Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.155713 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562315-j9xc6"] Mar 17 09:15:00 crc kubenswrapper[4813]: E0317 09:15:00.156831 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="404fc67e-83be-4be6-bf6e-4279c0543316" containerName="registry-server" Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.156854 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="404fc67e-83be-4be6-bf6e-4279c0543316" containerName="registry-server" Mar 17 09:15:00 crc kubenswrapper[4813]: E0317 09:15:00.156886 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="404fc67e-83be-4be6-bf6e-4279c0543316" containerName="extract-utilities" Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.156896 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="404fc67e-83be-4be6-bf6e-4279c0543316" containerName="extract-utilities" Mar 17 09:15:00 crc kubenswrapper[4813]: E0317 09:15:00.156912 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="404fc67e-83be-4be6-bf6e-4279c0543316" containerName="extract-content" Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.156925 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="404fc67e-83be-4be6-bf6e-4279c0543316" containerName="extract-content" Mar 17 09:15:00 crc kubenswrapper[4813]: E0317 09:15:00.156957 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e837d92-c48e-4b0f-a55a-40bdaea5d9f0" containerName="oauth-openshift" Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.156966 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e837d92-c48e-4b0f-a55a-40bdaea5d9f0" containerName="oauth-openshift" Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.157334 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="404fc67e-83be-4be6-bf6e-4279c0543316" containerName="registry-server" Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.157373 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e837d92-c48e-4b0f-a55a-40bdaea5d9f0" containerName="oauth-openshift" Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.161192 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562315-j9xc6" Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.166380 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.166698 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.183952 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562315-j9xc6"] Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.257509 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpnvm\" (UniqueName: \"kubernetes.io/projected/3f4060f9-13f7-4b31-bc96-db1d3698de1a-kube-api-access-cpnvm\") pod \"collect-profiles-29562315-j9xc6\" (UID: \"3f4060f9-13f7-4b31-bc96-db1d3698de1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562315-j9xc6" Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.257654 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3f4060f9-13f7-4b31-bc96-db1d3698de1a-secret-volume\") pod \"collect-profiles-29562315-j9xc6\" (UID: \"3f4060f9-13f7-4b31-bc96-db1d3698de1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562315-j9xc6" Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.257809 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f4060f9-13f7-4b31-bc96-db1d3698de1a-config-volume\") pod \"collect-profiles-29562315-j9xc6\" (UID: \"3f4060f9-13f7-4b31-bc96-db1d3698de1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562315-j9xc6" Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.359380 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3f4060f9-13f7-4b31-bc96-db1d3698de1a-secret-volume\") pod \"collect-profiles-29562315-j9xc6\" (UID: \"3f4060f9-13f7-4b31-bc96-db1d3698de1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562315-j9xc6" Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.359488 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f4060f9-13f7-4b31-bc96-db1d3698de1a-config-volume\") pod \"collect-profiles-29562315-j9xc6\" (UID: \"3f4060f9-13f7-4b31-bc96-db1d3698de1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562315-j9xc6" Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.359556 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpnvm\" (UniqueName: \"kubernetes.io/projected/3f4060f9-13f7-4b31-bc96-db1d3698de1a-kube-api-access-cpnvm\") pod \"collect-profiles-29562315-j9xc6\" (UID: \"3f4060f9-13f7-4b31-bc96-db1d3698de1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562315-j9xc6" Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.361769 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f4060f9-13f7-4b31-bc96-db1d3698de1a-config-volume\") pod \"collect-profiles-29562315-j9xc6\" (UID: \"3f4060f9-13f7-4b31-bc96-db1d3698de1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562315-j9xc6" Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.366026 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3f4060f9-13f7-4b31-bc96-db1d3698de1a-secret-volume\") pod \"collect-profiles-29562315-j9xc6\" (UID: \"3f4060f9-13f7-4b31-bc96-db1d3698de1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562315-j9xc6" Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.392817 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpnvm\" (UniqueName: \"kubernetes.io/projected/3f4060f9-13f7-4b31-bc96-db1d3698de1a-kube-api-access-cpnvm\") pod \"collect-profiles-29562315-j9xc6\" (UID: \"3f4060f9-13f7-4b31-bc96-db1d3698de1a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562315-j9xc6" Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.492404 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562315-j9xc6" Mar 17 09:15:00 crc kubenswrapper[4813]: I0317 09:15:00.742793 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e837d92-c48e-4b0f-a55a-40bdaea5d9f0" path="/var/lib/kubelet/pods/5e837d92-c48e-4b0f-a55a-40bdaea5d9f0/volumes" Mar 17 09:15:01 crc kubenswrapper[4813]: I0317 09:15:01.033407 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562315-j9xc6"] Mar 17 09:15:01 crc kubenswrapper[4813]: I0317 09:15:01.504338 4813 generic.go:334] "Generic (PLEG): container finished" podID="3f4060f9-13f7-4b31-bc96-db1d3698de1a" containerID="2fa5a95242f0483f2a8d0e8c9b8d64f9d9219802e44bd5d4f691be1ed406fe37" exitCode=0 Mar 17 09:15:01 crc kubenswrapper[4813]: I0317 09:15:01.504397 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562315-j9xc6" event={"ID":"3f4060f9-13f7-4b31-bc96-db1d3698de1a","Type":"ContainerDied","Data":"2fa5a95242f0483f2a8d0e8c9b8d64f9d9219802e44bd5d4f691be1ed406fe37"} Mar 17 09:15:01 crc kubenswrapper[4813]: I0317 09:15:01.504686 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562315-j9xc6" event={"ID":"3f4060f9-13f7-4b31-bc96-db1d3698de1a","Type":"ContainerStarted","Data":"a6d4003bf341065bb2f711d4a6267a6cd0dc6bc6215366fe119deaf4271e346f"} Mar 17 09:15:03 crc kubenswrapper[4813]: I0317 09:15:03.008683 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562315-j9xc6" Mar 17 09:15:03 crc kubenswrapper[4813]: I0317 09:15:03.094037 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cpnvm\" (UniqueName: \"kubernetes.io/projected/3f4060f9-13f7-4b31-bc96-db1d3698de1a-kube-api-access-cpnvm\") pod \"3f4060f9-13f7-4b31-bc96-db1d3698de1a\" (UID: \"3f4060f9-13f7-4b31-bc96-db1d3698de1a\") " Mar 17 09:15:03 crc kubenswrapper[4813]: I0317 09:15:03.094152 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3f4060f9-13f7-4b31-bc96-db1d3698de1a-secret-volume\") pod \"3f4060f9-13f7-4b31-bc96-db1d3698de1a\" (UID: \"3f4060f9-13f7-4b31-bc96-db1d3698de1a\") " Mar 17 09:15:03 crc kubenswrapper[4813]: I0317 09:15:03.094236 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f4060f9-13f7-4b31-bc96-db1d3698de1a-config-volume\") pod \"3f4060f9-13f7-4b31-bc96-db1d3698de1a\" (UID: \"3f4060f9-13f7-4b31-bc96-db1d3698de1a\") " Mar 17 09:15:03 crc kubenswrapper[4813]: I0317 09:15:03.095198 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f4060f9-13f7-4b31-bc96-db1d3698de1a-config-volume" (OuterVolumeSpecName: "config-volume") pod "3f4060f9-13f7-4b31-bc96-db1d3698de1a" (UID: "3f4060f9-13f7-4b31-bc96-db1d3698de1a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:15:03 crc kubenswrapper[4813]: I0317 09:15:03.103793 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f4060f9-13f7-4b31-bc96-db1d3698de1a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "3f4060f9-13f7-4b31-bc96-db1d3698de1a" (UID: "3f4060f9-13f7-4b31-bc96-db1d3698de1a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:15:03 crc kubenswrapper[4813]: I0317 09:15:03.103860 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f4060f9-13f7-4b31-bc96-db1d3698de1a-kube-api-access-cpnvm" (OuterVolumeSpecName: "kube-api-access-cpnvm") pod "3f4060f9-13f7-4b31-bc96-db1d3698de1a" (UID: "3f4060f9-13f7-4b31-bc96-db1d3698de1a"). InnerVolumeSpecName "kube-api-access-cpnvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:15:03 crc kubenswrapper[4813]: I0317 09:15:03.195557 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cpnvm\" (UniqueName: \"kubernetes.io/projected/3f4060f9-13f7-4b31-bc96-db1d3698de1a-kube-api-access-cpnvm\") on node \"crc\" DevicePath \"\"" Mar 17 09:15:03 crc kubenswrapper[4813]: I0317 09:15:03.196119 4813 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/3f4060f9-13f7-4b31-bc96-db1d3698de1a-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 17 09:15:03 crc kubenswrapper[4813]: I0317 09:15:03.196137 4813 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3f4060f9-13f7-4b31-bc96-db1d3698de1a-config-volume\") on node \"crc\" DevicePath \"\"" Mar 17 09:15:03 crc kubenswrapper[4813]: I0317 09:15:03.525047 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562315-j9xc6" event={"ID":"3f4060f9-13f7-4b31-bc96-db1d3698de1a","Type":"ContainerDied","Data":"a6d4003bf341065bb2f711d4a6267a6cd0dc6bc6215366fe119deaf4271e346f"} Mar 17 09:15:03 crc kubenswrapper[4813]: I0317 09:15:03.525125 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6d4003bf341065bb2f711d4a6267a6cd0dc6bc6215366fe119deaf4271e346f" Mar 17 09:15:03 crc kubenswrapper[4813]: I0317 09:15:03.525204 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562315-j9xc6" Mar 17 09:15:06 crc kubenswrapper[4813]: I0317 09:15:06.966121 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-756fdfd6b5-h422v"] Mar 17 09:15:06 crc kubenswrapper[4813]: E0317 09:15:06.966519 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f4060f9-13f7-4b31-bc96-db1d3698de1a" containerName="collect-profiles" Mar 17 09:15:06 crc kubenswrapper[4813]: I0317 09:15:06.966551 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f4060f9-13f7-4b31-bc96-db1d3698de1a" containerName="collect-profiles" Mar 17 09:15:06 crc kubenswrapper[4813]: I0317 09:15:06.966813 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f4060f9-13f7-4b31-bc96-db1d3698de1a" containerName="collect-profiles" Mar 17 09:15:06 crc kubenswrapper[4813]: I0317 09:15:06.967412 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:06 crc kubenswrapper[4813]: I0317 09:15:06.974109 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 17 09:15:06 crc kubenswrapper[4813]: I0317 09:15:06.974961 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 17 09:15:06 crc kubenswrapper[4813]: I0317 09:15:06.975193 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 17 09:15:06 crc kubenswrapper[4813]: I0317 09:15:06.975518 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 17 09:15:06 crc kubenswrapper[4813]: I0317 09:15:06.975749 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 17 09:15:06 crc kubenswrapper[4813]: I0317 09:15:06.975798 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 17 09:15:06 crc kubenswrapper[4813]: I0317 09:15:06.975761 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 17 09:15:06 crc kubenswrapper[4813]: I0317 09:15:06.977171 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 17 09:15:06 crc kubenswrapper[4813]: I0317 09:15:06.977938 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 17 09:15:06 crc kubenswrapper[4813]: I0317 09:15:06.978380 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 17 09:15:06 crc kubenswrapper[4813]: I0317 09:15:06.978588 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 17 09:15:06 crc kubenswrapper[4813]: I0317 09:15:06.978925 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 17 09:15:06 crc kubenswrapper[4813]: I0317 09:15:06.984693 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 17 09:15:06 crc kubenswrapper[4813]: I0317 09:15:06.999328 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.005889 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.030723 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-756fdfd6b5-h422v"] Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.046464 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.046514 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.046542 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-user-template-error\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.046584 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-system-router-certs\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.046658 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-user-template-login\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.046765 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-system-serving-cert\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.046830 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-system-session\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.046867 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.046894 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f71b792d-c12b-4087-9887-c2916d560bee-audit-policies\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.046911 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x58c2\" (UniqueName: \"kubernetes.io/projected/f71b792d-c12b-4087-9887-c2916d560bee-kube-api-access-x58c2\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.046928 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-system-service-ca\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.046999 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-system-cliconfig\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.047051 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f71b792d-c12b-4087-9887-c2916d560bee-audit-dir\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.047097 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.148810 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-system-router-certs\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.149103 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-user-template-login\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.149256 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-system-serving-cert\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.149358 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-system-session\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.149458 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.149555 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f71b792d-c12b-4087-9887-c2916d560bee-audit-policies\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.149663 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x58c2\" (UniqueName: \"kubernetes.io/projected/f71b792d-c12b-4087-9887-c2916d560bee-kube-api-access-x58c2\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.149768 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-system-service-ca\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.149862 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-system-cliconfig\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.149967 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f71b792d-c12b-4087-9887-c2916d560bee-audit-dir\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.150084 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.150160 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f71b792d-c12b-4087-9887-c2916d560bee-audit-dir\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.150271 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.150371 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.150451 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f71b792d-c12b-4087-9887-c2916d560bee-audit-policies\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.150469 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-user-template-error\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.151272 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.152944 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-system-cliconfig\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.154113 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-system-service-ca\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.154768 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.154796 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-system-session\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.156415 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-system-router-certs\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.156988 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-user-template-error\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.157152 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.159282 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-system-serving-cert\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.160406 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-user-template-login\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.163866 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f71b792d-c12b-4087-9887-c2916d560bee-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.172104 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x58c2\" (UniqueName: \"kubernetes.io/projected/f71b792d-c12b-4087-9887-c2916d560bee-kube-api-access-x58c2\") pod \"oauth-openshift-756fdfd6b5-h422v\" (UID: \"f71b792d-c12b-4087-9887-c2916d560bee\") " pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.283677 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:07 crc kubenswrapper[4813]: I0317 09:15:07.775859 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-756fdfd6b5-h422v"] Mar 17 09:15:08 crc kubenswrapper[4813]: I0317 09:15:08.565716 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" event={"ID":"f71b792d-c12b-4087-9887-c2916d560bee","Type":"ContainerStarted","Data":"3389117a16719ffd84bf5f35d9784c6b1e09842eedc333b3df75bc92a79c2e94"} Mar 17 09:15:08 crc kubenswrapper[4813]: I0317 09:15:08.565767 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" event={"ID":"f71b792d-c12b-4087-9887-c2916d560bee","Type":"ContainerStarted","Data":"bcf22f37e33dc264a5482e0863a9145ee093c807e7e0d58e70a2fbd97f268f9b"} Mar 17 09:15:08 crc kubenswrapper[4813]: I0317 09:15:08.566144 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:08 crc kubenswrapper[4813]: I0317 09:15:08.571942 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" Mar 17 09:15:08 crc kubenswrapper[4813]: I0317 09:15:08.597053 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-756fdfd6b5-h422v" podStartSLOduration=36.596715513 podStartE2EDuration="36.596715513s" podCreationTimestamp="2026-03-17 09:14:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:15:08.595641606 +0000 UTC m=+330.696445115" watchObservedRunningTime="2026-03-17 09:15:08.596715513 +0000 UTC m=+330.697519022" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.250910 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7549dfdb87-dkkng"] Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.251724 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" podUID="7207af26-0b14-4ce2-b178-2daad0226706" containerName="controller-manager" containerID="cri-o://6a9a555ebe60ef6ea5e88a305c1e1dd5899eb32ec794cb61aa4c77b8a8f77da3" gracePeriod=30 Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.343209 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8"] Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.343489 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" podUID="3261033c-f194-459b-952a-79391f6b3ea9" containerName="route-controller-manager" containerID="cri-o://c5c9eb13ac2c4ae5a424a8105d447b12e53d0dcfc5d76e9abba0ff72c354c798" gracePeriod=30 Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.581975 4813 generic.go:334] "Generic (PLEG): container finished" podID="7207af26-0b14-4ce2-b178-2daad0226706" containerID="6a9a555ebe60ef6ea5e88a305c1e1dd5899eb32ec794cb61aa4c77b8a8f77da3" exitCode=0 Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.582059 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" event={"ID":"7207af26-0b14-4ce2-b178-2daad0226706","Type":"ContainerDied","Data":"6a9a555ebe60ef6ea5e88a305c1e1dd5899eb32ec794cb61aa4c77b8a8f77da3"} Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.584190 4813 generic.go:334] "Generic (PLEG): container finished" podID="3261033c-f194-459b-952a-79391f6b3ea9" containerID="c5c9eb13ac2c4ae5a424a8105d447b12e53d0dcfc5d76e9abba0ff72c354c798" exitCode=0 Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.584302 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" event={"ID":"3261033c-f194-459b-952a-79391f6b3ea9","Type":"ContainerDied","Data":"c5c9eb13ac2c4ae5a424a8105d447b12e53d0dcfc5d76e9abba0ff72c354c798"} Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.589326 4813 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.589630 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab" gracePeriod=15 Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.589674 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281" gracePeriod=15 Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.589685 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://d76b00866bfb10cc5b5d6e623f6ae3e733020c40adf36dd1354bce3bddc64986" gracePeriod=15 Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.589738 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d" gracePeriod=15 Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.589974 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba" gracePeriod=15 Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.591076 4813 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 17 09:15:09 crc kubenswrapper[4813]: E0317 09:15:09.591387 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.591418 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 09:15:09 crc kubenswrapper[4813]: E0317 09:15:09.591432 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.591445 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 09:15:09 crc kubenswrapper[4813]: E0317 09:15:09.591462 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.591471 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 17 09:15:09 crc kubenswrapper[4813]: E0317 09:15:09.591483 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.591493 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 17 09:15:09 crc kubenswrapper[4813]: E0317 09:15:09.591506 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.591517 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 09:15:09 crc kubenswrapper[4813]: E0317 09:15:09.591528 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.591538 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 17 09:15:09 crc kubenswrapper[4813]: E0317 09:15:09.591552 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.591562 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 09:15:09 crc kubenswrapper[4813]: E0317 09:15:09.591574 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.591584 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 17 09:15:09 crc kubenswrapper[4813]: E0317 09:15:09.591624 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.591668 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 17 09:15:09 crc kubenswrapper[4813]: E0317 09:15:09.591684 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.591696 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.591838 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.591856 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.591870 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.591883 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.591894 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.591908 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.591920 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.591933 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.591946 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.591959 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 09:15:09 crc kubenswrapper[4813]: E0317 09:15:09.592227 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.592242 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.593920 4813 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.595929 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.600088 4813 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Mar 17 09:15:09 crc kubenswrapper[4813]: E0317 09:15:09.635051 4813 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.143:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.784079 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.784239 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.785065 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.785123 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.785210 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.785478 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.794671 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.795468 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.883783 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.884347 4813 status_manager.go:851] "Failed to get status for pod" podUID="3261033c-f194-459b-952a-79391f6b3ea9" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7767bbd8d-lw8t8\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.896043 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6w87z\" (UniqueName: \"kubernetes.io/projected/3261033c-f194-459b-952a-79391f6b3ea9-kube-api-access-6w87z\") pod \"3261033c-f194-459b-952a-79391f6b3ea9\" (UID: \"3261033c-f194-459b-952a-79391f6b3ea9\") " Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.896094 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3261033c-f194-459b-952a-79391f6b3ea9-client-ca\") pod \"3261033c-f194-459b-952a-79391f6b3ea9\" (UID: \"3261033c-f194-459b-952a-79391f6b3ea9\") " Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.896141 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3261033c-f194-459b-952a-79391f6b3ea9-serving-cert\") pod \"3261033c-f194-459b-952a-79391f6b3ea9\" (UID: \"3261033c-f194-459b-952a-79391f6b3ea9\") " Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.896161 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3261033c-f194-459b-952a-79391f6b3ea9-config\") pod \"3261033c-f194-459b-952a-79391f6b3ea9\" (UID: \"3261033c-f194-459b-952a-79391f6b3ea9\") " Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.896260 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.896287 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.896308 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.896322 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.896336 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.896366 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.896384 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.896420 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.896492 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.896513 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.896525 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.896970 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.897031 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.896981 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.897007 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.897026 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3261033c-f194-459b-952a-79391f6b3ea9-client-ca" (OuterVolumeSpecName: "client-ca") pod "3261033c-f194-459b-952a-79391f6b3ea9" (UID: "3261033c-f194-459b-952a-79391f6b3ea9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.896991 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.897546 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3261033c-f194-459b-952a-79391f6b3ea9-config" (OuterVolumeSpecName: "config") pod "3261033c-f194-459b-952a-79391f6b3ea9" (UID: "3261033c-f194-459b-952a-79391f6b3ea9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.901073 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3261033c-f194-459b-952a-79391f6b3ea9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "3261033c-f194-459b-952a-79391f6b3ea9" (UID: "3261033c-f194-459b-952a-79391f6b3ea9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.901315 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3261033c-f194-459b-952a-79391f6b3ea9-kube-api-access-6w87z" (OuterVolumeSpecName: "kube-api-access-6w87z") pod "3261033c-f194-459b-952a-79391f6b3ea9" (UID: "3261033c-f194-459b-952a-79391f6b3ea9"). InnerVolumeSpecName "kube-api-access-6w87z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.936694 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.949860 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.950435 4813 status_manager.go:851] "Failed to get status for pod" podUID="3261033c-f194-459b-952a-79391f6b3ea9" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7767bbd8d-lw8t8\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.950781 4813 status_manager.go:851] "Failed to get status for pod" podUID="7207af26-0b14-4ce2-b178-2daad0226706" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-7549dfdb87-dkkng\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:09 crc kubenswrapper[4813]: E0317 09:15:09.958653 4813 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.143:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189d9624144be55d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:15:09.957350749 +0000 UTC m=+332.058154248,LastTimestamp:2026-03-17 09:15:09.957350749 +0000 UTC m=+332.058154248,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.997385 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3261033c-f194-459b-952a-79391f6b3ea9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.997529 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3261033c-f194-459b-952a-79391f6b3ea9-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.997629 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6w87z\" (UniqueName: \"kubernetes.io/projected/3261033c-f194-459b-952a-79391f6b3ea9-kube-api-access-6w87z\") on node \"crc\" DevicePath \"\"" Mar 17 09:15:09 crc kubenswrapper[4813]: I0317 09:15:09.998269 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3261033c-f194-459b-952a-79391f6b3ea9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.098989 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7207af26-0b14-4ce2-b178-2daad0226706-config\") pod \"7207af26-0b14-4ce2-b178-2daad0226706\" (UID: \"7207af26-0b14-4ce2-b178-2daad0226706\") " Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.099054 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7207af26-0b14-4ce2-b178-2daad0226706-serving-cert\") pod \"7207af26-0b14-4ce2-b178-2daad0226706\" (UID: \"7207af26-0b14-4ce2-b178-2daad0226706\") " Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.099125 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7207af26-0b14-4ce2-b178-2daad0226706-client-ca\") pod \"7207af26-0b14-4ce2-b178-2daad0226706\" (UID: \"7207af26-0b14-4ce2-b178-2daad0226706\") " Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.100283 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7207af26-0b14-4ce2-b178-2daad0226706-proxy-ca-bundles\") pod \"7207af26-0b14-4ce2-b178-2daad0226706\" (UID: \"7207af26-0b14-4ce2-b178-2daad0226706\") " Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.100318 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5bgj\" (UniqueName: \"kubernetes.io/projected/7207af26-0b14-4ce2-b178-2daad0226706-kube-api-access-g5bgj\") pod \"7207af26-0b14-4ce2-b178-2daad0226706\" (UID: \"7207af26-0b14-4ce2-b178-2daad0226706\") " Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.099886 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7207af26-0b14-4ce2-b178-2daad0226706-client-ca" (OuterVolumeSpecName: "client-ca") pod "7207af26-0b14-4ce2-b178-2daad0226706" (UID: "7207af26-0b14-4ce2-b178-2daad0226706"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.100174 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7207af26-0b14-4ce2-b178-2daad0226706-config" (OuterVolumeSpecName: "config") pod "7207af26-0b14-4ce2-b178-2daad0226706" (UID: "7207af26-0b14-4ce2-b178-2daad0226706"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.100518 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7207af26-0b14-4ce2-b178-2daad0226706-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.100533 4813 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7207af26-0b14-4ce2-b178-2daad0226706-client-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.100746 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7207af26-0b14-4ce2-b178-2daad0226706-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7207af26-0b14-4ce2-b178-2daad0226706" (UID: "7207af26-0b14-4ce2-b178-2daad0226706"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.103567 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7207af26-0b14-4ce2-b178-2daad0226706-kube-api-access-g5bgj" (OuterVolumeSpecName: "kube-api-access-g5bgj") pod "7207af26-0b14-4ce2-b178-2daad0226706" (UID: "7207af26-0b14-4ce2-b178-2daad0226706"). InnerVolumeSpecName "kube-api-access-g5bgj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.104406 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7207af26-0b14-4ce2-b178-2daad0226706-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7207af26-0b14-4ce2-b178-2daad0226706" (UID: "7207af26-0b14-4ce2-b178-2daad0226706"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.201191 4813 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7207af26-0b14-4ce2-b178-2daad0226706-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.201241 4813 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7207af26-0b14-4ce2-b178-2daad0226706-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.201263 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5bgj\" (UniqueName: \"kubernetes.io/projected/7207af26-0b14-4ce2-b178-2daad0226706-kube-api-access-g5bgj\") on node \"crc\" DevicePath \"\"" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.596050 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/4.log" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.598861 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.600189 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d76b00866bfb10cc5b5d6e623f6ae3e733020c40adf36dd1354bce3bddc64986" exitCode=0 Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.600249 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba" exitCode=0 Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.600270 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281" exitCode=0 Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.600290 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d" exitCode=2 Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.600354 4813 scope.go:117] "RemoveContainer" containerID="05611f07591c02fc9de4b19ee837dd962cb578b61b0677ddc4cbd95a527efd2f" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.603944 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" event={"ID":"3261033c-f194-459b-952a-79391f6b3ea9","Type":"ContainerDied","Data":"3fb614813cba56cbb39aeb44d7a979eb10a9c9c30087a17b2ab0ad70d1c039fd"} Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.603997 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.605359 4813 status_manager.go:851] "Failed to get status for pod" podUID="7207af26-0b14-4ce2-b178-2daad0226706" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-7549dfdb87-dkkng\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.605923 4813 status_manager.go:851] "Failed to get status for pod" podUID="3261033c-f194-459b-952a-79391f6b3ea9" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7767bbd8d-lw8t8\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.608880 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"92e222d9d09a3aff09a08b4712d2d64c4908407c090ea6e41b62a3ad95e4b66a"} Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.608963 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"e4f29e7859462ce030a2218af24fd7fc624f3a95e0598ffa2c1e00fa36fe534e"} Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.609748 4813 status_manager.go:851] "Failed to get status for pod" podUID="3261033c-f194-459b-952a-79391f6b3ea9" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7767bbd8d-lw8t8\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:10 crc kubenswrapper[4813]: E0317 09:15:10.610007 4813 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.143:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.610288 4813 status_manager.go:851] "Failed to get status for pod" podUID="7207af26-0b14-4ce2-b178-2daad0226706" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-7549dfdb87-dkkng\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.612104 4813 generic.go:334] "Generic (PLEG): container finished" podID="e87d1b34-2547-4acd-8843-a8bffc737563" containerID="31bded1816cfd4a21a18bdf8b38e4e14c4cc7e06adde9dc74da11540e79e2499" exitCode=0 Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.612224 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e87d1b34-2547-4acd-8843-a8bffc737563","Type":"ContainerDied","Data":"31bded1816cfd4a21a18bdf8b38e4e14c4cc7e06adde9dc74da11540e79e2499"} Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.613010 4813 status_manager.go:851] "Failed to get status for pod" podUID="7207af26-0b14-4ce2-b178-2daad0226706" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-7549dfdb87-dkkng\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.613720 4813 status_manager.go:851] "Failed to get status for pod" podUID="3261033c-f194-459b-952a-79391f6b3ea9" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7767bbd8d-lw8t8\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.614095 4813 status_manager.go:851] "Failed to get status for pod" podUID="e87d1b34-2547-4acd-8843-a8bffc737563" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.615935 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.615937 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" event={"ID":"7207af26-0b14-4ce2-b178-2daad0226706","Type":"ContainerDied","Data":"cc2ab0d3a32b145182a81ceabe5cfba9debd88314f18399b00f8e068e5b28bee"} Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.620298 4813 status_manager.go:851] "Failed to get status for pod" podUID="e87d1b34-2547-4acd-8843-a8bffc737563" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.620761 4813 status_manager.go:851] "Failed to get status for pod" podUID="7207af26-0b14-4ce2-b178-2daad0226706" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-7549dfdb87-dkkng\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.621138 4813 status_manager.go:851] "Failed to get status for pod" podUID="3261033c-f194-459b-952a-79391f6b3ea9" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7767bbd8d-lw8t8\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.652314 4813 status_manager.go:851] "Failed to get status for pod" podUID="e87d1b34-2547-4acd-8843-a8bffc737563" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.653096 4813 status_manager.go:851] "Failed to get status for pod" podUID="7207af26-0b14-4ce2-b178-2daad0226706" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-7549dfdb87-dkkng\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.653524 4813 status_manager.go:851] "Failed to get status for pod" podUID="3261033c-f194-459b-952a-79391f6b3ea9" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7767bbd8d-lw8t8\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.653793 4813 status_manager.go:851] "Failed to get status for pod" podUID="e87d1b34-2547-4acd-8843-a8bffc737563" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.654139 4813 status_manager.go:851] "Failed to get status for pod" podUID="7207af26-0b14-4ce2-b178-2daad0226706" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-7549dfdb87-dkkng\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.655366 4813 status_manager.go:851] "Failed to get status for pod" podUID="3261033c-f194-459b-952a-79391f6b3ea9" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7767bbd8d-lw8t8\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.660020 4813 scope.go:117] "RemoveContainer" containerID="c5c9eb13ac2c4ae5a424a8105d447b12e53d0dcfc5d76e9abba0ff72c354c798" Mar 17 09:15:10 crc kubenswrapper[4813]: I0317 09:15:10.691624 4813 scope.go:117] "RemoveContainer" containerID="6a9a555ebe60ef6ea5e88a305c1e1dd5899eb32ec794cb61aa4c77b8a8f77da3" Mar 17 09:15:11 crc kubenswrapper[4813]: I0317 09:15:11.647569 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 17 09:15:11 crc kubenswrapper[4813]: I0317 09:15:11.957227 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 17 09:15:11 crc kubenswrapper[4813]: I0317 09:15:11.958839 4813 status_manager.go:851] "Failed to get status for pod" podUID="e87d1b34-2547-4acd-8843-a8bffc737563" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:11 crc kubenswrapper[4813]: I0317 09:15:11.959704 4813 status_manager.go:851] "Failed to get status for pod" podUID="7207af26-0b14-4ce2-b178-2daad0226706" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-7549dfdb87-dkkng\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:11 crc kubenswrapper[4813]: I0317 09:15:11.960197 4813 status_manager.go:851] "Failed to get status for pod" podUID="3261033c-f194-459b-952a-79391f6b3ea9" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7767bbd8d-lw8t8\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:11 crc kubenswrapper[4813]: I0317 09:15:11.963771 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 17 09:15:11 crc kubenswrapper[4813]: I0317 09:15:11.964591 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:11 crc kubenswrapper[4813]: I0317 09:15:11.965198 4813 status_manager.go:851] "Failed to get status for pod" podUID="7207af26-0b14-4ce2-b178-2daad0226706" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-7549dfdb87-dkkng\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:11 crc kubenswrapper[4813]: I0317 09:15:11.965522 4813 status_manager.go:851] "Failed to get status for pod" podUID="3261033c-f194-459b-952a-79391f6b3ea9" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7767bbd8d-lw8t8\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:11 crc kubenswrapper[4813]: I0317 09:15:11.965870 4813 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:11 crc kubenswrapper[4813]: I0317 09:15:11.966307 4813 status_manager.go:851] "Failed to get status for pod" podUID="e87d1b34-2547-4acd-8843-a8bffc737563" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.032034 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.032095 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.032139 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e87d1b34-2547-4acd-8843-a8bffc737563-kubelet-dir\") pod \"e87d1b34-2547-4acd-8843-a8bffc737563\" (UID: \"e87d1b34-2547-4acd-8843-a8bffc737563\") " Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.032177 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.032222 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.032296 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.032328 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e87d1b34-2547-4acd-8843-a8bffc737563-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e87d1b34-2547-4acd-8843-a8bffc737563" (UID: "e87d1b34-2547-4acd-8843-a8bffc737563"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.032339 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e87d1b34-2547-4acd-8843-a8bffc737563-kube-api-access\") pod \"e87d1b34-2547-4acd-8843-a8bffc737563\" (UID: \"e87d1b34-2547-4acd-8843-a8bffc737563\") " Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.032353 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.032376 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e87d1b34-2547-4acd-8843-a8bffc737563-var-lock\") pod \"e87d1b34-2547-4acd-8843-a8bffc737563\" (UID: \"e87d1b34-2547-4acd-8843-a8bffc737563\") " Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.032718 4813 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.032742 4813 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.032758 4813 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.032773 4813 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e87d1b34-2547-4acd-8843-a8bffc737563-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.032811 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e87d1b34-2547-4acd-8843-a8bffc737563-var-lock" (OuterVolumeSpecName: "var-lock") pod "e87d1b34-2547-4acd-8843-a8bffc737563" (UID: "e87d1b34-2547-4acd-8843-a8bffc737563"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.041057 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e87d1b34-2547-4acd-8843-a8bffc737563-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e87d1b34-2547-4acd-8843-a8bffc737563" (UID: "e87d1b34-2547-4acd-8843-a8bffc737563"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.133861 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e87d1b34-2547-4acd-8843-a8bffc737563-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.134273 4813 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/e87d1b34-2547-4acd-8843-a8bffc737563-var-lock\") on node \"crc\" DevicePath \"\"" Mar 17 09:15:12 crc kubenswrapper[4813]: E0317 09:15:12.456289 4813 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:12 crc kubenswrapper[4813]: E0317 09:15:12.457768 4813 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:12 crc kubenswrapper[4813]: E0317 09:15:12.458671 4813 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:12 crc kubenswrapper[4813]: E0317 09:15:12.458932 4813 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:12 crc kubenswrapper[4813]: E0317 09:15:12.459236 4813 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.459292 4813 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 17 09:15:12 crc kubenswrapper[4813]: E0317 09:15:12.459948 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.143:6443: connect: connection refused" interval="200ms" Mar 17 09:15:12 crc kubenswrapper[4813]: E0317 09:15:12.660934 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.143:6443: connect: connection refused" interval="400ms" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.676443 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.677146 4813 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab" exitCode=0 Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.677307 4813 scope.go:117] "RemoveContainer" containerID="d76b00866bfb10cc5b5d6e623f6ae3e733020c40adf36dd1354bce3bddc64986" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.677442 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.678917 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"e87d1b34-2547-4acd-8843-a8bffc737563","Type":"ContainerDied","Data":"e4bf54f7cff4843be1351bd95b6e5fd183e2b1043c2434596bd48b41348281ae"} Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.678976 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e4bf54f7cff4843be1351bd95b6e5fd183e2b1043c2434596bd48b41348281ae" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.679064 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.694151 4813 status_manager.go:851] "Failed to get status for pod" podUID="7207af26-0b14-4ce2-b178-2daad0226706" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-7549dfdb87-dkkng\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.694528 4813 status_manager.go:851] "Failed to get status for pod" podUID="3261033c-f194-459b-952a-79391f6b3ea9" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7767bbd8d-lw8t8\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.694767 4813 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.695263 4813 status_manager.go:851] "Failed to get status for pod" podUID="e87d1b34-2547-4acd-8843-a8bffc737563" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.695329 4813 scope.go:117] "RemoveContainer" containerID="ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.704893 4813 status_manager.go:851] "Failed to get status for pod" podUID="e87d1b34-2547-4acd-8843-a8bffc737563" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.705056 4813 status_manager.go:851] "Failed to get status for pod" podUID="7207af26-0b14-4ce2-b178-2daad0226706" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-7549dfdb87-dkkng\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.705201 4813 status_manager.go:851] "Failed to get status for pod" podUID="3261033c-f194-459b-952a-79391f6b3ea9" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7767bbd8d-lw8t8\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.705351 4813 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.709475 4813 scope.go:117] "RemoveContainer" containerID="3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.721786 4813 scope.go:117] "RemoveContainer" containerID="a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.736059 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.737223 4813 scope.go:117] "RemoveContainer" containerID="fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.754612 4813 scope.go:117] "RemoveContainer" containerID="9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.774857 4813 scope.go:117] "RemoveContainer" containerID="d76b00866bfb10cc5b5d6e623f6ae3e733020c40adf36dd1354bce3bddc64986" Mar 17 09:15:12 crc kubenswrapper[4813]: E0317 09:15:12.775338 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d76b00866bfb10cc5b5d6e623f6ae3e733020c40adf36dd1354bce3bddc64986\": container with ID starting with d76b00866bfb10cc5b5d6e623f6ae3e733020c40adf36dd1354bce3bddc64986 not found: ID does not exist" containerID="d76b00866bfb10cc5b5d6e623f6ae3e733020c40adf36dd1354bce3bddc64986" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.775378 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d76b00866bfb10cc5b5d6e623f6ae3e733020c40adf36dd1354bce3bddc64986"} err="failed to get container status \"d76b00866bfb10cc5b5d6e623f6ae3e733020c40adf36dd1354bce3bddc64986\": rpc error: code = NotFound desc = could not find container \"d76b00866bfb10cc5b5d6e623f6ae3e733020c40adf36dd1354bce3bddc64986\": container with ID starting with d76b00866bfb10cc5b5d6e623f6ae3e733020c40adf36dd1354bce3bddc64986 not found: ID does not exist" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.775410 4813 scope.go:117] "RemoveContainer" containerID="ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba" Mar 17 09:15:12 crc kubenswrapper[4813]: E0317 09:15:12.776211 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\": container with ID starting with ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba not found: ID does not exist" containerID="ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.776238 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba"} err="failed to get container status \"ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\": rpc error: code = NotFound desc = could not find container \"ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba\": container with ID starting with ec784df5da5efa718b3894bd702e9a8d4f980d43c67328eaa03fb07269b3eaba not found: ID does not exist" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.776255 4813 scope.go:117] "RemoveContainer" containerID="3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281" Mar 17 09:15:12 crc kubenswrapper[4813]: E0317 09:15:12.776580 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\": container with ID starting with 3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281 not found: ID does not exist" containerID="3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.776661 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281"} err="failed to get container status \"3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\": rpc error: code = NotFound desc = could not find container \"3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281\": container with ID starting with 3479828d9cdffd1d9f3f3d2df6cb03e43c574c9d8fa2d19faf2994b1f537f281 not found: ID does not exist" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.776680 4813 scope.go:117] "RemoveContainer" containerID="a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d" Mar 17 09:15:12 crc kubenswrapper[4813]: E0317 09:15:12.777208 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\": container with ID starting with a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d not found: ID does not exist" containerID="a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.777235 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d"} err="failed to get container status \"a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\": rpc error: code = NotFound desc = could not find container \"a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d\": container with ID starting with a905b271688c5f7f881b7958b7b843f9571277da0dd25617f08042f410836a4d not found: ID does not exist" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.777249 4813 scope.go:117] "RemoveContainer" containerID="fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab" Mar 17 09:15:12 crc kubenswrapper[4813]: E0317 09:15:12.777567 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\": container with ID starting with fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab not found: ID does not exist" containerID="fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.777587 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab"} err="failed to get container status \"fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\": rpc error: code = NotFound desc = could not find container \"fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab\": container with ID starting with fcbae6697b7e1391a88b8fa3aa9162ebc301206d00c726d9c4909447052a18ab not found: ID does not exist" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.777620 4813 scope.go:117] "RemoveContainer" containerID="9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb" Mar 17 09:15:12 crc kubenswrapper[4813]: E0317 09:15:12.777825 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\": container with ID starting with 9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb not found: ID does not exist" containerID="9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb" Mar 17 09:15:12 crc kubenswrapper[4813]: I0317 09:15:12.777849 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb"} err="failed to get container status \"9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\": rpc error: code = NotFound desc = could not find container \"9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb\": container with ID starting with 9d9eaaf26b65ddec327eb360d5398eae49b88b128a4730b3d051f5f20c1cc5eb not found: ID does not exist" Mar 17 09:15:13 crc kubenswrapper[4813]: E0317 09:15:13.062057 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.143:6443: connect: connection refused" interval="800ms" Mar 17 09:15:13 crc kubenswrapper[4813]: E0317 09:15:13.862871 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.143:6443: connect: connection refused" interval="1.6s" Mar 17 09:15:15 crc kubenswrapper[4813]: E0317 09:15:15.464794 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.143:6443: connect: connection refused" interval="3.2s" Mar 17 09:15:18 crc kubenswrapper[4813]: E0317 09:15:18.665994 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.143:6443: connect: connection refused" interval="6.4s" Mar 17 09:15:18 crc kubenswrapper[4813]: I0317 09:15:18.735340 4813 status_manager.go:851] "Failed to get status for pod" podUID="7207af26-0b14-4ce2-b178-2daad0226706" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-7549dfdb87-dkkng\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:18 crc kubenswrapper[4813]: I0317 09:15:18.735799 4813 status_manager.go:851] "Failed to get status for pod" podUID="3261033c-f194-459b-952a-79391f6b3ea9" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7767bbd8d-lw8t8\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:18 crc kubenswrapper[4813]: I0317 09:15:18.736384 4813 status_manager.go:851] "Failed to get status for pod" podUID="e87d1b34-2547-4acd-8843-a8bffc737563" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:19 crc kubenswrapper[4813]: E0317 09:15:19.606769 4813 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.143:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189d9624144be55d openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-17 09:15:09.957350749 +0000 UTC m=+332.058154248,LastTimestamp:2026-03-17 09:15:09.957350749 +0000 UTC m=+332.058154248,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 17 09:15:24 crc kubenswrapper[4813]: I0317 09:15:24.730664 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:24 crc kubenswrapper[4813]: I0317 09:15:24.732340 4813 status_manager.go:851] "Failed to get status for pod" podUID="3261033c-f194-459b-952a-79391f6b3ea9" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7767bbd8d-lw8t8\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:24 crc kubenswrapper[4813]: I0317 09:15:24.733266 4813 status_manager.go:851] "Failed to get status for pod" podUID="e87d1b34-2547-4acd-8843-a8bffc737563" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:24 crc kubenswrapper[4813]: I0317 09:15:24.737106 4813 status_manager.go:851] "Failed to get status for pod" podUID="7207af26-0b14-4ce2-b178-2daad0226706" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-7549dfdb87-dkkng\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:24 crc kubenswrapper[4813]: I0317 09:15:24.755150 4813 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a21775d2-fd48-430d-bd5c-679b7bb2c271" Mar 17 09:15:24 crc kubenswrapper[4813]: I0317 09:15:24.755212 4813 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a21775d2-fd48-430d-bd5c-679b7bb2c271" Mar 17 09:15:24 crc kubenswrapper[4813]: E0317 09:15:24.756028 4813 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.143:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:24 crc kubenswrapper[4813]: I0317 09:15:24.756883 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:24 crc kubenswrapper[4813]: I0317 09:15:24.777298 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 17 09:15:24 crc kubenswrapper[4813]: I0317 09:15:24.777998 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 17 09:15:24 crc kubenswrapper[4813]: I0317 09:15:24.778049 4813 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="ee1d7eaebd9e22a3f68b00c57f0bc9971826fc05a41062947a1ce04f1e8d380c" exitCode=1 Mar 17 09:15:24 crc kubenswrapper[4813]: I0317 09:15:24.778088 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"ee1d7eaebd9e22a3f68b00c57f0bc9971826fc05a41062947a1ce04f1e8d380c"} Mar 17 09:15:24 crc kubenswrapper[4813]: I0317 09:15:24.778700 4813 scope.go:117] "RemoveContainer" containerID="ee1d7eaebd9e22a3f68b00c57f0bc9971826fc05a41062947a1ce04f1e8d380c" Mar 17 09:15:24 crc kubenswrapper[4813]: I0317 09:15:24.780284 4813 status_manager.go:851] "Failed to get status for pod" podUID="e87d1b34-2547-4acd-8843-a8bffc737563" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:24 crc kubenswrapper[4813]: I0317 09:15:24.781009 4813 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:24 crc kubenswrapper[4813]: I0317 09:15:24.781672 4813 status_manager.go:851] "Failed to get status for pod" podUID="7207af26-0b14-4ce2-b178-2daad0226706" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-7549dfdb87-dkkng\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:24 crc kubenswrapper[4813]: I0317 09:15:24.782231 4813 status_manager.go:851] "Failed to get status for pod" podUID="3261033c-f194-459b-952a-79391f6b3ea9" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7767bbd8d-lw8t8\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:24 crc kubenswrapper[4813]: W0317 09:15:24.788949 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-c2a76fe709fabf147646cf68dbbd8cd6cc454aa12ea02bd474869107bfce7656 WatchSource:0}: Error finding container c2a76fe709fabf147646cf68dbbd8cd6cc454aa12ea02bd474869107bfce7656: Status 404 returned error can't find the container with id c2a76fe709fabf147646cf68dbbd8cd6cc454aa12ea02bd474869107bfce7656 Mar 17 09:15:25 crc kubenswrapper[4813]: E0317 09:15:25.067511 4813 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.143:6443: connect: connection refused" interval="7s" Mar 17 09:15:25 crc kubenswrapper[4813]: I0317 09:15:25.788821 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 17 09:15:25 crc kubenswrapper[4813]: I0317 09:15:25.789561 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 17 09:15:25 crc kubenswrapper[4813]: I0317 09:15:25.789727 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d4989f6db4b4507de0098bf32662c31928d1b8838593774a46c8c3b62175cb75"} Mar 17 09:15:25 crc kubenswrapper[4813]: I0317 09:15:25.790870 4813 status_manager.go:851] "Failed to get status for pod" podUID="3261033c-f194-459b-952a-79391f6b3ea9" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7767bbd8d-lw8t8\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:25 crc kubenswrapper[4813]: I0317 09:15:25.791359 4813 status_manager.go:851] "Failed to get status for pod" podUID="e87d1b34-2547-4acd-8843-a8bffc737563" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:25 crc kubenswrapper[4813]: I0317 09:15:25.791878 4813 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="4ae562a26f64786feaabe07a9ae323e2cacfd94824c5bce206f1053e466b99aa" exitCode=0 Mar 17 09:15:25 crc kubenswrapper[4813]: I0317 09:15:25.791905 4813 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:25 crc kubenswrapper[4813]: I0317 09:15:25.791935 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"4ae562a26f64786feaabe07a9ae323e2cacfd94824c5bce206f1053e466b99aa"} Mar 17 09:15:25 crc kubenswrapper[4813]: I0317 09:15:25.791982 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c2a76fe709fabf147646cf68dbbd8cd6cc454aa12ea02bd474869107bfce7656"} Mar 17 09:15:25 crc kubenswrapper[4813]: I0317 09:15:25.792390 4813 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a21775d2-fd48-430d-bd5c-679b7bb2c271" Mar 17 09:15:25 crc kubenswrapper[4813]: I0317 09:15:25.792374 4813 status_manager.go:851] "Failed to get status for pod" podUID="7207af26-0b14-4ce2-b178-2daad0226706" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-7549dfdb87-dkkng\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:25 crc kubenswrapper[4813]: I0317 09:15:25.792415 4813 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a21775d2-fd48-430d-bd5c-679b7bb2c271" Mar 17 09:15:25 crc kubenswrapper[4813]: I0317 09:15:25.793114 4813 status_manager.go:851] "Failed to get status for pod" podUID="3261033c-f194-459b-952a-79391f6b3ea9" pod="openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-route-controller-manager/pods/route-controller-manager-7767bbd8d-lw8t8\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:25 crc kubenswrapper[4813]: E0317 09:15:25.793208 4813 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.143:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:25 crc kubenswrapper[4813]: I0317 09:15:25.793625 4813 status_manager.go:851] "Failed to get status for pod" podUID="e87d1b34-2547-4acd-8843-a8bffc737563" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:25 crc kubenswrapper[4813]: I0317 09:15:25.794059 4813 status_manager.go:851] "Failed to get status for pod" podUID="7207af26-0b14-4ce2-b178-2daad0226706" pod="openshift-controller-manager/controller-manager-7549dfdb87-dkkng" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-controller-manager/pods/controller-manager-7549dfdb87-dkkng\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:25 crc kubenswrapper[4813]: I0317 09:15:25.794722 4813 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.143:6443: connect: connection refused" Mar 17 09:15:26 crc kubenswrapper[4813]: I0317 09:15:26.801434 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"a3e12f260288f6b17c77b0dab89abd97470a29bfb4f9572f6633dd212ea13545"} Mar 17 09:15:26 crc kubenswrapper[4813]: I0317 09:15:26.801755 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2298c73b4fb38756c95f9ab89590033b11bf95a774ad8926ef1fc13faeca9510"} Mar 17 09:15:26 crc kubenswrapper[4813]: I0317 09:15:26.801766 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b38e5fff529f74369bc966ece8ae476b33c909588cf55f12bff1c16d302f2093"} Mar 17 09:15:27 crc kubenswrapper[4813]: I0317 09:15:27.809099 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"1fea739c5eb485271ee0483eaa17ad0f0847cbf3592bf8a23dbc538dab067c0d"} Mar 17 09:15:27 crc kubenswrapper[4813]: I0317 09:15:27.809311 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"b84d7f36420e1ed2ec0e5bbeb65d463a0d8d3eec92d979db468426215706d6ba"} Mar 17 09:15:27 crc kubenswrapper[4813]: I0317 09:15:27.809392 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:27 crc kubenswrapper[4813]: I0317 09:15:27.809416 4813 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a21775d2-fd48-430d-bd5c-679b7bb2c271" Mar 17 09:15:27 crc kubenswrapper[4813]: I0317 09:15:27.809442 4813 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a21775d2-fd48-430d-bd5c-679b7bb2c271" Mar 17 09:15:29 crc kubenswrapper[4813]: I0317 09:15:29.756978 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:29 crc kubenswrapper[4813]: I0317 09:15:29.757834 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:29 crc kubenswrapper[4813]: I0317 09:15:29.766413 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:30 crc kubenswrapper[4813]: I0317 09:15:30.547528 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:15:30 crc kubenswrapper[4813]: I0317 09:15:30.552542 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:15:30 crc kubenswrapper[4813]: I0317 09:15:30.827689 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:15:32 crc kubenswrapper[4813]: I0317 09:15:32.820414 4813 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:32 crc kubenswrapper[4813]: I0317 09:15:32.840021 4813 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a21775d2-fd48-430d-bd5c-679b7bb2c271" Mar 17 09:15:32 crc kubenswrapper[4813]: I0317 09:15:32.840052 4813 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a21775d2-fd48-430d-bd5c-679b7bb2c271" Mar 17 09:15:32 crc kubenswrapper[4813]: I0317 09:15:32.844786 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:32 crc kubenswrapper[4813]: I0317 09:15:32.930885 4813 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="8848f847-4beb-44dc-9663-4de07d1138a5" Mar 17 09:15:33 crc kubenswrapper[4813]: I0317 09:15:33.846991 4813 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a21775d2-fd48-430d-bd5c-679b7bb2c271" Mar 17 09:15:33 crc kubenswrapper[4813]: I0317 09:15:33.847037 4813 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a21775d2-fd48-430d-bd5c-679b7bb2c271" Mar 17 09:15:33 crc kubenswrapper[4813]: I0317 09:15:33.853210 4813 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="8848f847-4beb-44dc-9663-4de07d1138a5" Mar 17 09:15:41 crc kubenswrapper[4813]: I0317 09:15:41.883510 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 17 09:15:42 crc kubenswrapper[4813]: I0317 09:15:42.094086 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 17 09:15:42 crc kubenswrapper[4813]: I0317 09:15:42.802933 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 17 09:15:43 crc kubenswrapper[4813]: I0317 09:15:43.297805 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 17 09:15:43 crc kubenswrapper[4813]: I0317 09:15:43.385714 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 17 09:15:43 crc kubenswrapper[4813]: I0317 09:15:43.545486 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 17 09:15:43 crc kubenswrapper[4813]: I0317 09:15:43.586839 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 17 09:15:43 crc kubenswrapper[4813]: I0317 09:15:43.769383 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 17 09:15:43 crc kubenswrapper[4813]: I0317 09:15:43.817824 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 17 09:15:43 crc kubenswrapper[4813]: I0317 09:15:43.819778 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 17 09:15:43 crc kubenswrapper[4813]: I0317 09:15:43.939205 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 17 09:15:44 crc kubenswrapper[4813]: I0317 09:15:44.060803 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 17 09:15:44 crc kubenswrapper[4813]: I0317 09:15:44.353245 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 17 09:15:44 crc kubenswrapper[4813]: I0317 09:15:44.570509 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 17 09:15:44 crc kubenswrapper[4813]: I0317 09:15:44.644093 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 17 09:15:44 crc kubenswrapper[4813]: I0317 09:15:44.647778 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 17 09:15:44 crc kubenswrapper[4813]: I0317 09:15:44.689872 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 17 09:15:44 crc kubenswrapper[4813]: I0317 09:15:44.699929 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 17 09:15:44 crc kubenswrapper[4813]: I0317 09:15:44.905733 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 17 09:15:44 crc kubenswrapper[4813]: I0317 09:15:44.956748 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 17 09:15:45 crc kubenswrapper[4813]: I0317 09:15:45.014142 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 17 09:15:45 crc kubenswrapper[4813]: I0317 09:15:45.028957 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 17 09:15:45 crc kubenswrapper[4813]: I0317 09:15:45.051086 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 17 09:15:45 crc kubenswrapper[4813]: I0317 09:15:45.080831 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 17 09:15:45 crc kubenswrapper[4813]: I0317 09:15:45.100077 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 17 09:15:45 crc kubenswrapper[4813]: I0317 09:15:45.238327 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 17 09:15:45 crc kubenswrapper[4813]: I0317 09:15:45.298025 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 17 09:15:45 crc kubenswrapper[4813]: I0317 09:15:45.349073 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 17 09:15:45 crc kubenswrapper[4813]: I0317 09:15:45.372215 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 17 09:15:45 crc kubenswrapper[4813]: I0317 09:15:45.394461 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 17 09:15:45 crc kubenswrapper[4813]: I0317 09:15:45.486648 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 17 09:15:45 crc kubenswrapper[4813]: I0317 09:15:45.516431 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 17 09:15:45 crc kubenswrapper[4813]: I0317 09:15:45.610980 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 17 09:15:45 crc kubenswrapper[4813]: I0317 09:15:45.611925 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 17 09:15:45 crc kubenswrapper[4813]: I0317 09:15:45.638810 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 17 09:15:45 crc kubenswrapper[4813]: I0317 09:15:45.639364 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 17 09:15:45 crc kubenswrapper[4813]: I0317 09:15:45.710069 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 17 09:15:45 crc kubenswrapper[4813]: I0317 09:15:45.756744 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 17 09:15:45 crc kubenswrapper[4813]: I0317 09:15:45.834930 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 17 09:15:45 crc kubenswrapper[4813]: I0317 09:15:45.957664 4813 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 17 09:15:45 crc kubenswrapper[4813]: I0317 09:15:45.975888 4813 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 17 09:15:46 crc kubenswrapper[4813]: I0317 09:15:46.010441 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 17 09:15:46 crc kubenswrapper[4813]: I0317 09:15:46.115445 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 17 09:15:46 crc kubenswrapper[4813]: I0317 09:15:46.223486 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 17 09:15:46 crc kubenswrapper[4813]: I0317 09:15:46.241189 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 17 09:15:46 crc kubenswrapper[4813]: I0317 09:15:46.253017 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 17 09:15:46 crc kubenswrapper[4813]: I0317 09:15:46.616650 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 17 09:15:46 crc kubenswrapper[4813]: I0317 09:15:46.642981 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 17 09:15:46 crc kubenswrapper[4813]: I0317 09:15:46.663499 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 17 09:15:46 crc kubenswrapper[4813]: I0317 09:15:46.756001 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 17 09:15:46 crc kubenswrapper[4813]: I0317 09:15:46.855335 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 17 09:15:47 crc kubenswrapper[4813]: I0317 09:15:47.007438 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 17 09:15:47 crc kubenswrapper[4813]: I0317 09:15:47.046666 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 17 09:15:47 crc kubenswrapper[4813]: I0317 09:15:47.093532 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 17 09:15:47 crc kubenswrapper[4813]: I0317 09:15:47.162713 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 17 09:15:47 crc kubenswrapper[4813]: I0317 09:15:47.192496 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 17 09:15:47 crc kubenswrapper[4813]: I0317 09:15:47.206273 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 17 09:15:47 crc kubenswrapper[4813]: I0317 09:15:47.270278 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 17 09:15:47 crc kubenswrapper[4813]: I0317 09:15:47.496155 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 17 09:15:47 crc kubenswrapper[4813]: I0317 09:15:47.603008 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 17 09:15:47 crc kubenswrapper[4813]: I0317 09:15:47.648135 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 17 09:15:47 crc kubenswrapper[4813]: I0317 09:15:47.827053 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 17 09:15:47 crc kubenswrapper[4813]: I0317 09:15:47.839208 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 17 09:15:47 crc kubenswrapper[4813]: I0317 09:15:47.894725 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 17 09:15:47 crc kubenswrapper[4813]: I0317 09:15:47.903208 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 17 09:15:48 crc kubenswrapper[4813]: I0317 09:15:48.002699 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 17 09:15:48 crc kubenswrapper[4813]: I0317 09:15:48.130984 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 17 09:15:48 crc kubenswrapper[4813]: I0317 09:15:48.207791 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 17 09:15:48 crc kubenswrapper[4813]: I0317 09:15:48.256725 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 17 09:15:48 crc kubenswrapper[4813]: I0317 09:15:48.509886 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 17 09:15:48 crc kubenswrapper[4813]: I0317 09:15:48.517302 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 17 09:15:48 crc kubenswrapper[4813]: I0317 09:15:48.549848 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 17 09:15:48 crc kubenswrapper[4813]: I0317 09:15:48.688299 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 17 09:15:48 crc kubenswrapper[4813]: I0317 09:15:48.904001 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 17 09:15:48 crc kubenswrapper[4813]: I0317 09:15:48.909463 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 17 09:15:48 crc kubenswrapper[4813]: I0317 09:15:48.915246 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.018302 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.031819 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.181330 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.208212 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.225887 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.238713 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.309679 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.314069 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.412776 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.419571 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.440424 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.529919 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.540114 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.548397 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.558654 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.577875 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.599503 4813 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.663170 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.691631 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.740655 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.775009 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.780827 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.841374 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 17 09:15:49 crc kubenswrapper[4813]: I0317 09:15:49.915630 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 17 09:15:50 crc kubenswrapper[4813]: I0317 09:15:50.065284 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 17 09:15:50 crc kubenswrapper[4813]: I0317 09:15:50.369654 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 17 09:15:50 crc kubenswrapper[4813]: I0317 09:15:50.425316 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 17 09:15:50 crc kubenswrapper[4813]: I0317 09:15:50.505692 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 17 09:15:50 crc kubenswrapper[4813]: I0317 09:15:50.524938 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 17 09:15:50 crc kubenswrapper[4813]: I0317 09:15:50.546421 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 17 09:15:50 crc kubenswrapper[4813]: I0317 09:15:50.612159 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 17 09:15:50 crc kubenswrapper[4813]: I0317 09:15:50.680141 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 17 09:15:50 crc kubenswrapper[4813]: I0317 09:15:50.763582 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 17 09:15:50 crc kubenswrapper[4813]: I0317 09:15:50.805816 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 17 09:15:50 crc kubenswrapper[4813]: I0317 09:15:50.807941 4813 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 17 09:15:50 crc kubenswrapper[4813]: I0317 09:15:50.853949 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 17 09:15:50 crc kubenswrapper[4813]: I0317 09:15:50.928719 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 17 09:15:51 crc kubenswrapper[4813]: I0317 09:15:51.178347 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 17 09:15:51 crc kubenswrapper[4813]: I0317 09:15:51.181852 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 17 09:15:51 crc kubenswrapper[4813]: I0317 09:15:51.208090 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 17 09:15:51 crc kubenswrapper[4813]: I0317 09:15:51.317749 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 17 09:15:51 crc kubenswrapper[4813]: I0317 09:15:51.334924 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 17 09:15:51 crc kubenswrapper[4813]: I0317 09:15:51.338479 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 17 09:15:51 crc kubenswrapper[4813]: I0317 09:15:51.359219 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 17 09:15:51 crc kubenswrapper[4813]: I0317 09:15:51.377964 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 17 09:15:51 crc kubenswrapper[4813]: I0317 09:15:51.382855 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 17 09:15:51 crc kubenswrapper[4813]: I0317 09:15:51.440440 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 17 09:15:51 crc kubenswrapper[4813]: I0317 09:15:51.453462 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 17 09:15:51 crc kubenswrapper[4813]: I0317 09:15:51.618799 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 17 09:15:51 crc kubenswrapper[4813]: I0317 09:15:51.643742 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 17 09:15:51 crc kubenswrapper[4813]: I0317 09:15:51.662492 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 17 09:15:51 crc kubenswrapper[4813]: I0317 09:15:51.693887 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 17 09:15:51 crc kubenswrapper[4813]: I0317 09:15:51.693981 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 17 09:15:51 crc kubenswrapper[4813]: I0317 09:15:51.714541 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 17 09:15:51 crc kubenswrapper[4813]: I0317 09:15:51.872523 4813 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 17 09:15:51 crc kubenswrapper[4813]: I0317 09:15:51.962706 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 17 09:15:52 crc kubenswrapper[4813]: I0317 09:15:52.002285 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 17 09:15:52 crc kubenswrapper[4813]: I0317 09:15:52.007355 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 17 09:15:52 crc kubenswrapper[4813]: I0317 09:15:52.047750 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 17 09:15:52 crc kubenswrapper[4813]: I0317 09:15:52.058077 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 17 09:15:52 crc kubenswrapper[4813]: I0317 09:15:52.074897 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 17 09:15:52 crc kubenswrapper[4813]: I0317 09:15:52.213080 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 17 09:15:52 crc kubenswrapper[4813]: I0317 09:15:52.290299 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 17 09:15:52 crc kubenswrapper[4813]: I0317 09:15:52.397324 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 17 09:15:52 crc kubenswrapper[4813]: I0317 09:15:52.422525 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 17 09:15:52 crc kubenswrapper[4813]: I0317 09:15:52.505497 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 17 09:15:52 crc kubenswrapper[4813]: I0317 09:15:52.557410 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 17 09:15:52 crc kubenswrapper[4813]: I0317 09:15:52.566068 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 17 09:15:52 crc kubenswrapper[4813]: I0317 09:15:52.598448 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 17 09:15:52 crc kubenswrapper[4813]: I0317 09:15:52.632548 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 17 09:15:52 crc kubenswrapper[4813]: I0317 09:15:52.687724 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 17 09:15:52 crc kubenswrapper[4813]: I0317 09:15:52.767182 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 17 09:15:52 crc kubenswrapper[4813]: I0317 09:15:52.848370 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 17 09:15:52 crc kubenswrapper[4813]: I0317 09:15:52.870245 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 17 09:15:53 crc kubenswrapper[4813]: I0317 09:15:53.057684 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 17 09:15:53 crc kubenswrapper[4813]: I0317 09:15:53.085011 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 17 09:15:53 crc kubenswrapper[4813]: I0317 09:15:53.243168 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 17 09:15:53 crc kubenswrapper[4813]: I0317 09:15:53.272644 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 17 09:15:53 crc kubenswrapper[4813]: I0317 09:15:53.281840 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 17 09:15:53 crc kubenswrapper[4813]: I0317 09:15:53.471268 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 17 09:15:53 crc kubenswrapper[4813]: I0317 09:15:53.474881 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 17 09:15:53 crc kubenswrapper[4813]: I0317 09:15:53.495584 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 17 09:15:53 crc kubenswrapper[4813]: I0317 09:15:53.514129 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 17 09:15:53 crc kubenswrapper[4813]: I0317 09:15:53.520543 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 17 09:15:53 crc kubenswrapper[4813]: I0317 09:15:53.539284 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 17 09:15:53 crc kubenswrapper[4813]: I0317 09:15:53.618696 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 17 09:15:53 crc kubenswrapper[4813]: I0317 09:15:53.719665 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 17 09:15:53 crc kubenswrapper[4813]: I0317 09:15:53.764557 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 17 09:15:53 crc kubenswrapper[4813]: I0317 09:15:53.816138 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 17 09:15:53 crc kubenswrapper[4813]: I0317 09:15:53.827404 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 17 09:15:53 crc kubenswrapper[4813]: I0317 09:15:53.846646 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 17 09:15:53 crc kubenswrapper[4813]: I0317 09:15:53.850274 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 17 09:15:53 crc kubenswrapper[4813]: I0317 09:15:53.898843 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 17 09:15:53 crc kubenswrapper[4813]: I0317 09:15:53.962028 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 17 09:15:53 crc kubenswrapper[4813]: I0317 09:15:53.974684 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 17 09:15:54 crc kubenswrapper[4813]: I0317 09:15:54.084804 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 17 09:15:54 crc kubenswrapper[4813]: I0317 09:15:54.155508 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 17 09:15:54 crc kubenswrapper[4813]: I0317 09:15:54.219685 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 17 09:15:54 crc kubenswrapper[4813]: I0317 09:15:54.221166 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 17 09:15:54 crc kubenswrapper[4813]: I0317 09:15:54.265436 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 17 09:15:54 crc kubenswrapper[4813]: I0317 09:15:54.430296 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 17 09:15:54 crc kubenswrapper[4813]: I0317 09:15:54.441863 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 17 09:15:54 crc kubenswrapper[4813]: I0317 09:15:54.471695 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 17 09:15:54 crc kubenswrapper[4813]: I0317 09:15:54.627408 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 17 09:15:54 crc kubenswrapper[4813]: I0317 09:15:54.687211 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 17 09:15:54 crc kubenswrapper[4813]: I0317 09:15:54.763629 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 17 09:15:54 crc kubenswrapper[4813]: I0317 09:15:54.797130 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 17 09:15:55 crc kubenswrapper[4813]: I0317 09:15:55.016949 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 17 09:15:55 crc kubenswrapper[4813]: I0317 09:15:55.109249 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 17 09:15:55 crc kubenswrapper[4813]: I0317 09:15:55.222393 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 17 09:15:55 crc kubenswrapper[4813]: I0317 09:15:55.233372 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 17 09:15:55 crc kubenswrapper[4813]: I0317 09:15:55.263766 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 17 09:15:55 crc kubenswrapper[4813]: I0317 09:15:55.373478 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 17 09:15:55 crc kubenswrapper[4813]: I0317 09:15:55.394903 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 17 09:15:55 crc kubenswrapper[4813]: I0317 09:15:55.454538 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 17 09:15:55 crc kubenswrapper[4813]: I0317 09:15:55.521905 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 17 09:15:55 crc kubenswrapper[4813]: I0317 09:15:55.557942 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 17 09:15:55 crc kubenswrapper[4813]: I0317 09:15:55.689369 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 17 09:15:55 crc kubenswrapper[4813]: I0317 09:15:55.697245 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 17 09:15:55 crc kubenswrapper[4813]: I0317 09:15:55.758008 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 17 09:15:55 crc kubenswrapper[4813]: I0317 09:15:55.777887 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 17 09:15:55 crc kubenswrapper[4813]: I0317 09:15:55.778982 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 17 09:15:55 crc kubenswrapper[4813]: I0317 09:15:55.781316 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 17 09:15:55 crc kubenswrapper[4813]: I0317 09:15:55.815107 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 17 09:15:55 crc kubenswrapper[4813]: I0317 09:15:55.887074 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 17 09:15:55 crc kubenswrapper[4813]: I0317 09:15:55.905428 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 17 09:15:55 crc kubenswrapper[4813]: I0317 09:15:55.923280 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 17 09:15:55 crc kubenswrapper[4813]: I0317 09:15:55.959395 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.021148 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.027311 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.136044 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.257951 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.392523 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.431556 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.436350 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.466824 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.577729 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.638873 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.907826 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.965000 4813 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.971443 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7767bbd8d-lw8t8","openshift-controller-manager/controller-manager-7549dfdb87-dkkng","openshift-kube-apiserver/kube-apiserver-crc"] Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.971515 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-controller-manager/controller-manager-85df77647c-g5jqs","openshift-route-controller-manager/route-controller-manager-56c59797f4-sdbq5"] Mar 17 09:15:56 crc kubenswrapper[4813]: E0317 09:15:56.971749 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3261033c-f194-459b-952a-79391f6b3ea9" containerName="route-controller-manager" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.971767 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3261033c-f194-459b-952a-79391f6b3ea9" containerName="route-controller-manager" Mar 17 09:15:56 crc kubenswrapper[4813]: E0317 09:15:56.971784 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e87d1b34-2547-4acd-8843-a8bffc737563" containerName="installer" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.971793 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e87d1b34-2547-4acd-8843-a8bffc737563" containerName="installer" Mar 17 09:15:56 crc kubenswrapper[4813]: E0317 09:15:56.971812 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7207af26-0b14-4ce2-b178-2daad0226706" containerName="controller-manager" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.971820 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7207af26-0b14-4ce2-b178-2daad0226706" containerName="controller-manager" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.971937 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="3261033c-f194-459b-952a-79391f6b3ea9" containerName="route-controller-manager" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.971953 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e87d1b34-2547-4acd-8843-a8bffc737563" containerName="installer" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.971967 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="7207af26-0b14-4ce2-b178-2daad0226706" containerName="controller-manager" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.972935 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.972994 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56c59797f4-sdbq5" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.973071 4813 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a21775d2-fd48-430d-bd5c-679b7bb2c271" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.973101 4813 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="a21775d2-fd48-430d-bd5c-679b7bb2c271" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.976847 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.976869 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.977130 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.977635 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.978951 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.979066 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.979141 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.979497 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.979498 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.979571 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.981282 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.981405 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.981579 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 17 09:15:56 crc kubenswrapper[4813]: I0317 09:15:56.990986 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.024558 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6f8a21b0-85c6-4312-a742-34c4cd29a74b-client-ca\") pod \"controller-manager-85df77647c-g5jqs\" (UID: \"6f8a21b0-85c6-4312-a742-34c4cd29a74b\") " pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.024670 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6f8a21b0-85c6-4312-a742-34c4cd29a74b-proxy-ca-bundles\") pod \"controller-manager-85df77647c-g5jqs\" (UID: \"6f8a21b0-85c6-4312-a742-34c4cd29a74b\") " pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.024715 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8a54d4c-573e-4fd9-8e94-bd651e3cec5a-config\") pod \"route-controller-manager-56c59797f4-sdbq5\" (UID: \"f8a54d4c-573e-4fd9-8e94-bd651e3cec5a\") " pod="openshift-route-controller-manager/route-controller-manager-56c59797f4-sdbq5" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.024744 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2bq9\" (UniqueName: \"kubernetes.io/projected/6f8a21b0-85c6-4312-a742-34c4cd29a74b-kube-api-access-g2bq9\") pod \"controller-manager-85df77647c-g5jqs\" (UID: \"6f8a21b0-85c6-4312-a742-34c4cd29a74b\") " pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.024806 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f8a21b0-85c6-4312-a742-34c4cd29a74b-serving-cert\") pod \"controller-manager-85df77647c-g5jqs\" (UID: \"6f8a21b0-85c6-4312-a742-34c4cd29a74b\") " pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.024853 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdbjf\" (UniqueName: \"kubernetes.io/projected/f8a54d4c-573e-4fd9-8e94-bd651e3cec5a-kube-api-access-xdbjf\") pod \"route-controller-manager-56c59797f4-sdbq5\" (UID: \"f8a54d4c-573e-4fd9-8e94-bd651e3cec5a\") " pod="openshift-route-controller-manager/route-controller-manager-56c59797f4-sdbq5" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.024957 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f8a54d4c-573e-4fd9-8e94-bd651e3cec5a-client-ca\") pod \"route-controller-manager-56c59797f4-sdbq5\" (UID: \"f8a54d4c-573e-4fd9-8e94-bd651e3cec5a\") " pod="openshift-route-controller-manager/route-controller-manager-56c59797f4-sdbq5" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.024997 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f8a21b0-85c6-4312-a742-34c4cd29a74b-config\") pod \"controller-manager-85df77647c-g5jqs\" (UID: \"6f8a21b0-85c6-4312-a742-34c4cd29a74b\") " pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.025088 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8a54d4c-573e-4fd9-8e94-bd651e3cec5a-serving-cert\") pod \"route-controller-manager-56c59797f4-sdbq5\" (UID: \"f8a54d4c-573e-4fd9-8e94-bd651e3cec5a\") " pod="openshift-route-controller-manager/route-controller-manager-56c59797f4-sdbq5" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.026939 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=25.026922702 podStartE2EDuration="25.026922702s" podCreationTimestamp="2026-03-17 09:15:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:15:57.025961478 +0000 UTC m=+379.126765007" watchObservedRunningTime="2026-03-17 09:15:57.026922702 +0000 UTC m=+379.127726211" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.118425 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.126402 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8a54d4c-573e-4fd9-8e94-bd651e3cec5a-serving-cert\") pod \"route-controller-manager-56c59797f4-sdbq5\" (UID: \"f8a54d4c-573e-4fd9-8e94-bd651e3cec5a\") " pod="openshift-route-controller-manager/route-controller-manager-56c59797f4-sdbq5" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.126460 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6f8a21b0-85c6-4312-a742-34c4cd29a74b-client-ca\") pod \"controller-manager-85df77647c-g5jqs\" (UID: \"6f8a21b0-85c6-4312-a742-34c4cd29a74b\") " pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.126492 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6f8a21b0-85c6-4312-a742-34c4cd29a74b-proxy-ca-bundles\") pod \"controller-manager-85df77647c-g5jqs\" (UID: \"6f8a21b0-85c6-4312-a742-34c4cd29a74b\") " pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.126519 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8a54d4c-573e-4fd9-8e94-bd651e3cec5a-config\") pod \"route-controller-manager-56c59797f4-sdbq5\" (UID: \"f8a54d4c-573e-4fd9-8e94-bd651e3cec5a\") " pod="openshift-route-controller-manager/route-controller-manager-56c59797f4-sdbq5" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.126538 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2bq9\" (UniqueName: \"kubernetes.io/projected/6f8a21b0-85c6-4312-a742-34c4cd29a74b-kube-api-access-g2bq9\") pod \"controller-manager-85df77647c-g5jqs\" (UID: \"6f8a21b0-85c6-4312-a742-34c4cd29a74b\") " pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.126575 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f8a21b0-85c6-4312-a742-34c4cd29a74b-serving-cert\") pod \"controller-manager-85df77647c-g5jqs\" (UID: \"6f8a21b0-85c6-4312-a742-34c4cd29a74b\") " pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.126614 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdbjf\" (UniqueName: \"kubernetes.io/projected/f8a54d4c-573e-4fd9-8e94-bd651e3cec5a-kube-api-access-xdbjf\") pod \"route-controller-manager-56c59797f4-sdbq5\" (UID: \"f8a54d4c-573e-4fd9-8e94-bd651e3cec5a\") " pod="openshift-route-controller-manager/route-controller-manager-56c59797f4-sdbq5" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.126640 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f8a54d4c-573e-4fd9-8e94-bd651e3cec5a-client-ca\") pod \"route-controller-manager-56c59797f4-sdbq5\" (UID: \"f8a54d4c-573e-4fd9-8e94-bd651e3cec5a\") " pod="openshift-route-controller-manager/route-controller-manager-56c59797f4-sdbq5" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.126659 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f8a21b0-85c6-4312-a742-34c4cd29a74b-config\") pod \"controller-manager-85df77647c-g5jqs\" (UID: \"6f8a21b0-85c6-4312-a742-34c4cd29a74b\") " pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.127357 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6f8a21b0-85c6-4312-a742-34c4cd29a74b-client-ca\") pod \"controller-manager-85df77647c-g5jqs\" (UID: \"6f8a21b0-85c6-4312-a742-34c4cd29a74b\") " pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.127901 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6f8a21b0-85c6-4312-a742-34c4cd29a74b-proxy-ca-bundles\") pod \"controller-manager-85df77647c-g5jqs\" (UID: \"6f8a21b0-85c6-4312-a742-34c4cd29a74b\") " pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.128408 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6f8a21b0-85c6-4312-a742-34c4cd29a74b-config\") pod \"controller-manager-85df77647c-g5jqs\" (UID: \"6f8a21b0-85c6-4312-a742-34c4cd29a74b\") " pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.128431 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f8a54d4c-573e-4fd9-8e94-bd651e3cec5a-client-ca\") pod \"route-controller-manager-56c59797f4-sdbq5\" (UID: \"f8a54d4c-573e-4fd9-8e94-bd651e3cec5a\") " pod="openshift-route-controller-manager/route-controller-manager-56c59797f4-sdbq5" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.133661 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8a54d4c-573e-4fd9-8e94-bd651e3cec5a-config\") pod \"route-controller-manager-56c59797f4-sdbq5\" (UID: \"f8a54d4c-573e-4fd9-8e94-bd651e3cec5a\") " pod="openshift-route-controller-manager/route-controller-manager-56c59797f4-sdbq5" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.146474 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6f8a21b0-85c6-4312-a742-34c4cd29a74b-serving-cert\") pod \"controller-manager-85df77647c-g5jqs\" (UID: \"6f8a21b0-85c6-4312-a742-34c4cd29a74b\") " pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.147095 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f8a54d4c-573e-4fd9-8e94-bd651e3cec5a-serving-cert\") pod \"route-controller-manager-56c59797f4-sdbq5\" (UID: \"f8a54d4c-573e-4fd9-8e94-bd651e3cec5a\") " pod="openshift-route-controller-manager/route-controller-manager-56c59797f4-sdbq5" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.149645 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2bq9\" (UniqueName: \"kubernetes.io/projected/6f8a21b0-85c6-4312-a742-34c4cd29a74b-kube-api-access-g2bq9\") pod \"controller-manager-85df77647c-g5jqs\" (UID: \"6f8a21b0-85c6-4312-a742-34c4cd29a74b\") " pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.155218 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdbjf\" (UniqueName: \"kubernetes.io/projected/f8a54d4c-573e-4fd9-8e94-bd651e3cec5a-kube-api-access-xdbjf\") pod \"route-controller-manager-56c59797f4-sdbq5\" (UID: \"f8a54d4c-573e-4fd9-8e94-bd651e3cec5a\") " pod="openshift-route-controller-manager/route-controller-manager-56c59797f4-sdbq5" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.831253 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.834105 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.836325 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.837492 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-56c59797f4-sdbq5" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.839550 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.839740 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.849100 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.852556 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.856958 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 17 09:15:57 crc kubenswrapper[4813]: I0317 09:15:57.903965 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 17 09:15:58 crc kubenswrapper[4813]: I0317 09:15:58.133186 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 17 09:15:58 crc kubenswrapper[4813]: I0317 09:15:58.341101 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-56c59797f4-sdbq5"] Mar 17 09:15:58 crc kubenswrapper[4813]: I0317 09:15:58.351042 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-85df77647c-g5jqs"] Mar 17 09:15:58 crc kubenswrapper[4813]: W0317 09:15:58.354299 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6f8a21b0_85c6_4312_a742_34c4cd29a74b.slice/crio-ec858e00089d516992fcce1f4d9f380a65c5b63a677996ed65ea1fa05c67ea57 WatchSource:0}: Error finding container ec858e00089d516992fcce1f4d9f380a65c5b63a677996ed65ea1fa05c67ea57: Status 404 returned error can't find the container with id ec858e00089d516992fcce1f4d9f380a65c5b63a677996ed65ea1fa05c67ea57 Mar 17 09:15:58 crc kubenswrapper[4813]: W0317 09:15:58.356232 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf8a54d4c_573e_4fd9_8e94_bd651e3cec5a.slice/crio-f9befed50467d1620ef9ac6cf4bc28c9dcf5d27718eea535347ef0ef52cde85a WatchSource:0}: Error finding container f9befed50467d1620ef9ac6cf4bc28c9dcf5d27718eea535347ef0ef52cde85a: Status 404 returned error can't find the container with id f9befed50467d1620ef9ac6cf4bc28c9dcf5d27718eea535347ef0ef52cde85a Mar 17 09:15:58 crc kubenswrapper[4813]: I0317 09:15:58.740993 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3261033c-f194-459b-952a-79391f6b3ea9" path="/var/lib/kubelet/pods/3261033c-f194-459b-952a-79391f6b3ea9/volumes" Mar 17 09:15:58 crc kubenswrapper[4813]: I0317 09:15:58.742156 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7207af26-0b14-4ce2-b178-2daad0226706" path="/var/lib/kubelet/pods/7207af26-0b14-4ce2-b178-2daad0226706/volumes" Mar 17 09:15:58 crc kubenswrapper[4813]: I0317 09:15:58.803330 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 17 09:15:59 crc kubenswrapper[4813]: I0317 09:15:59.020066 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" event={"ID":"6f8a21b0-85c6-4312-a742-34c4cd29a74b","Type":"ContainerStarted","Data":"24f4c1cc0dd748bf486075b43b8212c2e84c176758c5885a8e9c21ae9f076631"} Mar 17 09:15:59 crc kubenswrapper[4813]: I0317 09:15:59.020117 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" event={"ID":"6f8a21b0-85c6-4312-a742-34c4cd29a74b","Type":"ContainerStarted","Data":"ec858e00089d516992fcce1f4d9f380a65c5b63a677996ed65ea1fa05c67ea57"} Mar 17 09:15:59 crc kubenswrapper[4813]: I0317 09:15:59.021145 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" Mar 17 09:15:59 crc kubenswrapper[4813]: I0317 09:15:59.022703 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56c59797f4-sdbq5" event={"ID":"f8a54d4c-573e-4fd9-8e94-bd651e3cec5a","Type":"ContainerStarted","Data":"10a924e459c8add7571270794ad52c03e8b55ea2f595aef057af7274d11abb86"} Mar 17 09:15:59 crc kubenswrapper[4813]: I0317 09:15:59.022750 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-56c59797f4-sdbq5" event={"ID":"f8a54d4c-573e-4fd9-8e94-bd651e3cec5a","Type":"ContainerStarted","Data":"f9befed50467d1620ef9ac6cf4bc28c9dcf5d27718eea535347ef0ef52cde85a"} Mar 17 09:15:59 crc kubenswrapper[4813]: I0317 09:15:59.022913 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-56c59797f4-sdbq5" Mar 17 09:15:59 crc kubenswrapper[4813]: I0317 09:15:59.025544 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" Mar 17 09:15:59 crc kubenswrapper[4813]: I0317 09:15:59.054419 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 17 09:15:59 crc kubenswrapper[4813]: I0317 09:15:59.093234 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-85df77647c-g5jqs" podStartSLOduration=50.093217426 podStartE2EDuration="50.093217426s" podCreationTimestamp="2026-03-17 09:15:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:15:59.092889008 +0000 UTC m=+381.193692507" watchObservedRunningTime="2026-03-17 09:15:59.093217426 +0000 UTC m=+381.194020925" Mar 17 09:15:59 crc kubenswrapper[4813]: I0317 09:15:59.116210 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-56c59797f4-sdbq5" podStartSLOduration=50.116192743 podStartE2EDuration="50.116192743s" podCreationTimestamp="2026-03-17 09:15:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:15:59.114610684 +0000 UTC m=+381.215414183" watchObservedRunningTime="2026-03-17 09:15:59.116192743 +0000 UTC m=+381.216996242" Mar 17 09:15:59 crc kubenswrapper[4813]: I0317 09:15:59.163110 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 17 09:15:59 crc kubenswrapper[4813]: I0317 09:15:59.189733 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-56c59797f4-sdbq5" Mar 17 09:16:00 crc kubenswrapper[4813]: I0317 09:16:00.171074 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562316-rzldf"] Mar 17 09:16:00 crc kubenswrapper[4813]: I0317 09:16:00.171835 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562316-rzldf" Mar 17 09:16:00 crc kubenswrapper[4813]: I0317 09:16:00.174717 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:16:00 crc kubenswrapper[4813]: I0317 09:16:00.176223 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:16:00 crc kubenswrapper[4813]: I0317 09:16:00.176282 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:16:00 crc kubenswrapper[4813]: I0317 09:16:00.179782 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562316-rzldf"] Mar 17 09:16:00 crc kubenswrapper[4813]: I0317 09:16:00.266214 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8g27\" (UniqueName: \"kubernetes.io/projected/86b710cc-c717-4373-8b26-89d1d6d66bd6-kube-api-access-m8g27\") pod \"auto-csr-approver-29562316-rzldf\" (UID: \"86b710cc-c717-4373-8b26-89d1d6d66bd6\") " pod="openshift-infra/auto-csr-approver-29562316-rzldf" Mar 17 09:16:00 crc kubenswrapper[4813]: I0317 09:16:00.367162 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8g27\" (UniqueName: \"kubernetes.io/projected/86b710cc-c717-4373-8b26-89d1d6d66bd6-kube-api-access-m8g27\") pod \"auto-csr-approver-29562316-rzldf\" (UID: \"86b710cc-c717-4373-8b26-89d1d6d66bd6\") " pod="openshift-infra/auto-csr-approver-29562316-rzldf" Mar 17 09:16:00 crc kubenswrapper[4813]: I0317 09:16:00.401052 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8g27\" (UniqueName: \"kubernetes.io/projected/86b710cc-c717-4373-8b26-89d1d6d66bd6-kube-api-access-m8g27\") pod \"auto-csr-approver-29562316-rzldf\" (UID: \"86b710cc-c717-4373-8b26-89d1d6d66bd6\") " pod="openshift-infra/auto-csr-approver-29562316-rzldf" Mar 17 09:16:00 crc kubenswrapper[4813]: I0317 09:16:00.490807 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562316-rzldf" Mar 17 09:16:00 crc kubenswrapper[4813]: I0317 09:16:00.898512 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562316-rzldf"] Mar 17 09:16:00 crc kubenswrapper[4813]: W0317 09:16:00.908201 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod86b710cc_c717_4373_8b26_89d1d6d66bd6.slice/crio-2ff9691377d94a0e3dfb14b77ab5f5e82d110c582cf97bbe4c65e0d948a497f8 WatchSource:0}: Error finding container 2ff9691377d94a0e3dfb14b77ab5f5e82d110c582cf97bbe4c65e0d948a497f8: Status 404 returned error can't find the container with id 2ff9691377d94a0e3dfb14b77ab5f5e82d110c582cf97bbe4c65e0d948a497f8 Mar 17 09:16:01 crc kubenswrapper[4813]: I0317 09:16:01.043253 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562316-rzldf" event={"ID":"86b710cc-c717-4373-8b26-89d1d6d66bd6","Type":"ContainerStarted","Data":"2ff9691377d94a0e3dfb14b77ab5f5e82d110c582cf97bbe4c65e0d948a497f8"} Mar 17 09:16:03 crc kubenswrapper[4813]: I0317 09:16:03.079089 4813 generic.go:334] "Generic (PLEG): container finished" podID="86b710cc-c717-4373-8b26-89d1d6d66bd6" containerID="7ccdaaae72fb5d598178037112459ee1588a87f04acdb9b6eb7c0cfcfa485397" exitCode=0 Mar 17 09:16:03 crc kubenswrapper[4813]: I0317 09:16:03.079182 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562316-rzldf" event={"ID":"86b710cc-c717-4373-8b26-89d1d6d66bd6","Type":"ContainerDied","Data":"7ccdaaae72fb5d598178037112459ee1588a87f04acdb9b6eb7c0cfcfa485397"} Mar 17 09:16:04 crc kubenswrapper[4813]: I0317 09:16:04.488478 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562316-rzldf" Mar 17 09:16:04 crc kubenswrapper[4813]: I0317 09:16:04.622084 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m8g27\" (UniqueName: \"kubernetes.io/projected/86b710cc-c717-4373-8b26-89d1d6d66bd6-kube-api-access-m8g27\") pod \"86b710cc-c717-4373-8b26-89d1d6d66bd6\" (UID: \"86b710cc-c717-4373-8b26-89d1d6d66bd6\") " Mar 17 09:16:04 crc kubenswrapper[4813]: I0317 09:16:04.628906 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/86b710cc-c717-4373-8b26-89d1d6d66bd6-kube-api-access-m8g27" (OuterVolumeSpecName: "kube-api-access-m8g27") pod "86b710cc-c717-4373-8b26-89d1d6d66bd6" (UID: "86b710cc-c717-4373-8b26-89d1d6d66bd6"). InnerVolumeSpecName "kube-api-access-m8g27". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:16:04 crc kubenswrapper[4813]: I0317 09:16:04.723644 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m8g27\" (UniqueName: \"kubernetes.io/projected/86b710cc-c717-4373-8b26-89d1d6d66bd6-kube-api-access-m8g27\") on node \"crc\" DevicePath \"\"" Mar 17 09:16:05 crc kubenswrapper[4813]: I0317 09:16:05.097666 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562316-rzldf" event={"ID":"86b710cc-c717-4373-8b26-89d1d6d66bd6","Type":"ContainerDied","Data":"2ff9691377d94a0e3dfb14b77ab5f5e82d110c582cf97bbe4c65e0d948a497f8"} Mar 17 09:16:05 crc kubenswrapper[4813]: I0317 09:16:05.097735 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2ff9691377d94a0e3dfb14b77ab5f5e82d110c582cf97bbe4c65e0d948a497f8" Mar 17 09:16:05 crc kubenswrapper[4813]: I0317 09:16:05.097747 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562316-rzldf" Mar 17 09:16:06 crc kubenswrapper[4813]: I0317 09:16:06.308690 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 17 09:16:06 crc kubenswrapper[4813]: I0317 09:16:06.663099 4813 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 17 09:16:06 crc kubenswrapper[4813]: I0317 09:16:06.664336 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://92e222d9d09a3aff09a08b4712d2d64c4908407c090ea6e41b62a3ad95e4b66a" gracePeriod=5 Mar 17 09:16:08 crc kubenswrapper[4813]: I0317 09:16:08.812805 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 17 09:16:09 crc kubenswrapper[4813]: I0317 09:16:09.030633 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 17 09:16:12 crc kubenswrapper[4813]: I0317 09:16:12.135454 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 17 09:16:12 crc kubenswrapper[4813]: I0317 09:16:12.135936 4813 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="92e222d9d09a3aff09a08b4712d2d64c4908407c090ea6e41b62a3ad95e4b66a" exitCode=137 Mar 17 09:16:12 crc kubenswrapper[4813]: I0317 09:16:12.246552 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 17 09:16:12 crc kubenswrapper[4813]: I0317 09:16:12.246637 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 09:16:12 crc kubenswrapper[4813]: I0317 09:16:12.411021 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 17 09:16:12 crc kubenswrapper[4813]: I0317 09:16:12.411149 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 17 09:16:12 crc kubenswrapper[4813]: I0317 09:16:12.411259 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 17 09:16:12 crc kubenswrapper[4813]: I0317 09:16:12.411358 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 17 09:16:12 crc kubenswrapper[4813]: I0317 09:16:12.411399 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 17 09:16:12 crc kubenswrapper[4813]: I0317 09:16:12.411825 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:16:12 crc kubenswrapper[4813]: I0317 09:16:12.411839 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:16:12 crc kubenswrapper[4813]: I0317 09:16:12.411865 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:16:12 crc kubenswrapper[4813]: I0317 09:16:12.411850 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:16:12 crc kubenswrapper[4813]: I0317 09:16:12.423096 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:16:12 crc kubenswrapper[4813]: I0317 09:16:12.512669 4813 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 17 09:16:12 crc kubenswrapper[4813]: I0317 09:16:12.513010 4813 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 17 09:16:12 crc kubenswrapper[4813]: I0317 09:16:12.513029 4813 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 17 09:16:12 crc kubenswrapper[4813]: I0317 09:16:12.513043 4813 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 17 09:16:12 crc kubenswrapper[4813]: I0317 09:16:12.513058 4813 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 17 09:16:12 crc kubenswrapper[4813]: I0317 09:16:12.738802 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 17 09:16:13 crc kubenswrapper[4813]: I0317 09:16:13.143541 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 17 09:16:13 crc kubenswrapper[4813]: I0317 09:16:13.143665 4813 scope.go:117] "RemoveContainer" containerID="92e222d9d09a3aff09a08b4712d2d64c4908407c090ea6e41b62a3ad95e4b66a" Mar 17 09:16:13 crc kubenswrapper[4813]: I0317 09:16:13.143691 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 17 09:16:14 crc kubenswrapper[4813]: I0317 09:16:14.475423 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 17 09:16:18 crc kubenswrapper[4813]: I0317 09:16:18.267976 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 17 09:16:21 crc kubenswrapper[4813]: I0317 09:16:21.202410 4813 generic.go:334] "Generic (PLEG): container finished" podID="74f92156-f803-433a-99a4-981202076e81" containerID="f6ab163fea4b95b91341d3e6ebaee0ae1e7c34b859dfe7cfc8c1d4b4c0e83595" exitCode=0 Mar 17 09:16:21 crc kubenswrapper[4813]: I0317 09:16:21.202559 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" event={"ID":"74f92156-f803-433a-99a4-981202076e81","Type":"ContainerDied","Data":"f6ab163fea4b95b91341d3e6ebaee0ae1e7c34b859dfe7cfc8c1d4b4c0e83595"} Mar 17 09:16:21 crc kubenswrapper[4813]: I0317 09:16:21.203762 4813 scope.go:117] "RemoveContainer" containerID="f6ab163fea4b95b91341d3e6ebaee0ae1e7c34b859dfe7cfc8c1d4b4c0e83595" Mar 17 09:16:22 crc kubenswrapper[4813]: I0317 09:16:22.215256 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" event={"ID":"74f92156-f803-433a-99a4-981202076e81","Type":"ContainerStarted","Data":"a2ff2e083beeae6c658a5db2db106c161d0759072bf21050e7653c63d2bf67d5"} Mar 17 09:16:22 crc kubenswrapper[4813]: I0317 09:16:22.218843 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" Mar 17 09:16:22 crc kubenswrapper[4813]: I0317 09:16:22.221573 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" Mar 17 09:16:22 crc kubenswrapper[4813]: I0317 09:16:22.406865 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 17 09:16:26 crc kubenswrapper[4813]: I0317 09:16:26.065317 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 17 09:16:30 crc kubenswrapper[4813]: I0317 09:16:30.304260 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 17 09:16:32 crc kubenswrapper[4813]: I0317 09:16:32.282075 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 17 09:16:33 crc kubenswrapper[4813]: I0317 09:16:33.326103 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 17 09:16:35 crc kubenswrapper[4813]: I0317 09:16:35.199894 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 17 09:17:14 crc kubenswrapper[4813]: I0317 09:17:14.113753 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:17:14 crc kubenswrapper[4813]: I0317 09:17:14.114557 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.287982 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-m59rk"] Mar 17 09:17:16 crc kubenswrapper[4813]: E0317 09:17:16.288826 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.288861 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 17 09:17:16 crc kubenswrapper[4813]: E0317 09:17:16.288878 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="86b710cc-c717-4373-8b26-89d1d6d66bd6" containerName="oc" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.288892 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="86b710cc-c717-4373-8b26-89d1d6d66bd6" containerName="oc" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.289071 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.289102 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="86b710cc-c717-4373-8b26-89d1d6d66bd6" containerName="oc" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.289719 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.294406 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-m59rk"] Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.418514 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2c14d80e-5523-40fe-b046-8939f8e194fe-ca-trust-extracted\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.418574 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2c14d80e-5523-40fe-b046-8939f8e194fe-installation-pull-secrets\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.418695 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2c14d80e-5523-40fe-b046-8939f8e194fe-bound-sa-token\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.418745 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6kv5\" (UniqueName: \"kubernetes.io/projected/2c14d80e-5523-40fe-b046-8939f8e194fe-kube-api-access-v6kv5\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.418790 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2c14d80e-5523-40fe-b046-8939f8e194fe-trusted-ca\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.418835 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.418860 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2c14d80e-5523-40fe-b046-8939f8e194fe-registry-certificates\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.418883 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2c14d80e-5523-40fe-b046-8939f8e194fe-registry-tls\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.441003 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.520541 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6kv5\" (UniqueName: \"kubernetes.io/projected/2c14d80e-5523-40fe-b046-8939f8e194fe-kube-api-access-v6kv5\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.520627 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2c14d80e-5523-40fe-b046-8939f8e194fe-trusted-ca\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.520659 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2c14d80e-5523-40fe-b046-8939f8e194fe-registry-certificates\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.520682 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2c14d80e-5523-40fe-b046-8939f8e194fe-registry-tls\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.520738 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2c14d80e-5523-40fe-b046-8939f8e194fe-ca-trust-extracted\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.520777 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2c14d80e-5523-40fe-b046-8939f8e194fe-installation-pull-secrets\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.520853 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2c14d80e-5523-40fe-b046-8939f8e194fe-bound-sa-token\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.522151 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2c14d80e-5523-40fe-b046-8939f8e194fe-ca-trust-extracted\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.522886 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2c14d80e-5523-40fe-b046-8939f8e194fe-trusted-ca\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.523798 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2c14d80e-5523-40fe-b046-8939f8e194fe-registry-certificates\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.531110 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2c14d80e-5523-40fe-b046-8939f8e194fe-installation-pull-secrets\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.531112 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2c14d80e-5523-40fe-b046-8939f8e194fe-registry-tls\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.547862 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6kv5\" (UniqueName: \"kubernetes.io/projected/2c14d80e-5523-40fe-b046-8939f8e194fe-kube-api-access-v6kv5\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.548179 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2c14d80e-5523-40fe-b046-8939f8e194fe-bound-sa-token\") pod \"image-registry-66df7c8f76-m59rk\" (UID: \"2c14d80e-5523-40fe-b046-8939f8e194fe\") " pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.613591 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:16 crc kubenswrapper[4813]: I0317 09:17:16.880008 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-m59rk"] Mar 17 09:17:16 crc kubenswrapper[4813]: W0317 09:17:16.884174 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2c14d80e_5523_40fe_b046_8939f8e194fe.slice/crio-05961864ba85b8a50e4e22105fde89cd84e97e02ff66bed87915dbbee02da68b WatchSource:0}: Error finding container 05961864ba85b8a50e4e22105fde89cd84e97e02ff66bed87915dbbee02da68b: Status 404 returned error can't find the container with id 05961864ba85b8a50e4e22105fde89cd84e97e02ff66bed87915dbbee02da68b Mar 17 09:17:17 crc kubenswrapper[4813]: I0317 09:17:17.593232 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" event={"ID":"2c14d80e-5523-40fe-b046-8939f8e194fe","Type":"ContainerStarted","Data":"5094b38d720dad4b7a865ef4a99b365015bc9c08e281b8c1f5f907cc8abaa332"} Mar 17 09:17:17 crc kubenswrapper[4813]: I0317 09:17:17.593578 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:17 crc kubenswrapper[4813]: I0317 09:17:17.593625 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" event={"ID":"2c14d80e-5523-40fe-b046-8939f8e194fe","Type":"ContainerStarted","Data":"05961864ba85b8a50e4e22105fde89cd84e97e02ff66bed87915dbbee02da68b"} Mar 17 09:17:30 crc kubenswrapper[4813]: I0317 09:17:30.791062 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" podStartSLOduration=14.791045382 podStartE2EDuration="14.791045382s" podCreationTimestamp="2026-03-17 09:17:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:17:17.644339014 +0000 UTC m=+459.745142563" watchObservedRunningTime="2026-03-17 09:17:30.791045382 +0000 UTC m=+472.891848881" Mar 17 09:17:30 crc kubenswrapper[4813]: I0317 09:17:30.804002 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2skjk"] Mar 17 09:17:30 crc kubenswrapper[4813]: I0317 09:17:30.804263 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2skjk" podUID="65af6b35-41bd-4ad2-b8d4-0b653cb18dcf" containerName="registry-server" containerID="cri-o://e61a91f1f051970c805f6541cb76afeeb6855918e43560a4f856e2e03dcf12d9" gracePeriod=30 Mar 17 09:17:30 crc kubenswrapper[4813]: I0317 09:17:30.807586 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lxrc7"] Mar 17 09:17:30 crc kubenswrapper[4813]: I0317 09:17:30.807885 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-lxrc7" podUID="1b9e1de0-da0f-422e-a79b-4fb54d2e3acb" containerName="registry-server" containerID="cri-o://76af5a3320222301f14aee13eb616c8cc0bfc67f5aca3d2f0701ca64340366b2" gracePeriod=30 Mar 17 09:17:30 crc kubenswrapper[4813]: I0317 09:17:30.824500 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5wl9g"] Mar 17 09:17:30 crc kubenswrapper[4813]: I0317 09:17:30.825373 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" podUID="74f92156-f803-433a-99a4-981202076e81" containerName="marketplace-operator" containerID="cri-o://a2ff2e083beeae6c658a5db2db106c161d0759072bf21050e7653c63d2bf67d5" gracePeriod=30 Mar 17 09:17:30 crc kubenswrapper[4813]: I0317 09:17:30.831505 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-js422"] Mar 17 09:17:30 crc kubenswrapper[4813]: I0317 09:17:30.831740 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-js422" podUID="12cf87e0-5d1a-4598-bd14-e3ef92a56dc1" containerName="registry-server" containerID="cri-o://4e2f9e427afe63058f581dbffd89d35960f8c1034239db65888db4a7b6558688" gracePeriod=30 Mar 17 09:17:30 crc kubenswrapper[4813]: I0317 09:17:30.843118 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nw44l"] Mar 17 09:17:30 crc kubenswrapper[4813]: I0317 09:17:30.843969 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nw44l" Mar 17 09:17:30 crc kubenswrapper[4813]: I0317 09:17:30.849767 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6gvsx"] Mar 17 09:17:30 crc kubenswrapper[4813]: I0317 09:17:30.850127 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6gvsx" podUID="6b3a014a-3574-4c2b-8331-df326d47f240" containerName="registry-server" containerID="cri-o://3ccf2869040560c186883f68381aa6c02ceb23a455a78065f18335e2e2b78217" gracePeriod=30 Mar 17 09:17:30 crc kubenswrapper[4813]: I0317 09:17:30.853218 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nw44l"] Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.015194 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdph2\" (UniqueName: \"kubernetes.io/projected/6ca7f195-b5c8-4804-b95a-5ff53a637ac5-kube-api-access-jdph2\") pod \"marketplace-operator-79b997595-nw44l\" (UID: \"6ca7f195-b5c8-4804-b95a-5ff53a637ac5\") " pod="openshift-marketplace/marketplace-operator-79b997595-nw44l" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.015561 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6ca7f195-b5c8-4804-b95a-5ff53a637ac5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nw44l\" (UID: \"6ca7f195-b5c8-4804-b95a-5ff53a637ac5\") " pod="openshift-marketplace/marketplace-operator-79b997595-nw44l" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.015589 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6ca7f195-b5c8-4804-b95a-5ff53a637ac5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nw44l\" (UID: \"6ca7f195-b5c8-4804-b95a-5ff53a637ac5\") " pod="openshift-marketplace/marketplace-operator-79b997595-nw44l" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.116687 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6ca7f195-b5c8-4804-b95a-5ff53a637ac5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nw44l\" (UID: \"6ca7f195-b5c8-4804-b95a-5ff53a637ac5\") " pod="openshift-marketplace/marketplace-operator-79b997595-nw44l" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.116754 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6ca7f195-b5c8-4804-b95a-5ff53a637ac5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nw44l\" (UID: \"6ca7f195-b5c8-4804-b95a-5ff53a637ac5\") " pod="openshift-marketplace/marketplace-operator-79b997595-nw44l" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.116823 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdph2\" (UniqueName: \"kubernetes.io/projected/6ca7f195-b5c8-4804-b95a-5ff53a637ac5-kube-api-access-jdph2\") pod \"marketplace-operator-79b997595-nw44l\" (UID: \"6ca7f195-b5c8-4804-b95a-5ff53a637ac5\") " pod="openshift-marketplace/marketplace-operator-79b997595-nw44l" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.122481 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6ca7f195-b5c8-4804-b95a-5ff53a637ac5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-nw44l\" (UID: \"6ca7f195-b5c8-4804-b95a-5ff53a637ac5\") " pod="openshift-marketplace/marketplace-operator-79b997595-nw44l" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.124509 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/6ca7f195-b5c8-4804-b95a-5ff53a637ac5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-nw44l\" (UID: \"6ca7f195-b5c8-4804-b95a-5ff53a637ac5\") " pod="openshift-marketplace/marketplace-operator-79b997595-nw44l" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.135910 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdph2\" (UniqueName: \"kubernetes.io/projected/6ca7f195-b5c8-4804-b95a-5ff53a637ac5-kube-api-access-jdph2\") pod \"marketplace-operator-79b997595-nw44l\" (UID: \"6ca7f195-b5c8-4804-b95a-5ff53a637ac5\") " pod="openshift-marketplace/marketplace-operator-79b997595-nw44l" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.343472 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-nw44l" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.347680 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2skjk" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.354073 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lxrc7" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.363331 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.371086 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6gvsx" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.395152 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-js422" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.520742 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q8hk6\" (UniqueName: \"kubernetes.io/projected/74f92156-f803-433a-99a4-981202076e81-kube-api-access-q8hk6\") pod \"74f92156-f803-433a-99a4-981202076e81\" (UID: \"74f92156-f803-433a-99a4-981202076e81\") " Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.521104 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b3a014a-3574-4c2b-8331-df326d47f240-catalog-content\") pod \"6b3a014a-3574-4c2b-8331-df326d47f240\" (UID: \"6b3a014a-3574-4c2b-8331-df326d47f240\") " Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.521138 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12cf87e0-5d1a-4598-bd14-e3ef92a56dc1-utilities\") pod \"12cf87e0-5d1a-4598-bd14-e3ef92a56dc1\" (UID: \"12cf87e0-5d1a-4598-bd14-e3ef92a56dc1\") " Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.521166 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tt5rs\" (UniqueName: \"kubernetes.io/projected/1b9e1de0-da0f-422e-a79b-4fb54d2e3acb-kube-api-access-tt5rs\") pod \"1b9e1de0-da0f-422e-a79b-4fb54d2e3acb\" (UID: \"1b9e1de0-da0f-422e-a79b-4fb54d2e3acb\") " Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.521194 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6pmb\" (UniqueName: \"kubernetes.io/projected/65af6b35-41bd-4ad2-b8d4-0b653cb18dcf-kube-api-access-s6pmb\") pod \"65af6b35-41bd-4ad2-b8d4-0b653cb18dcf\" (UID: \"65af6b35-41bd-4ad2-b8d4-0b653cb18dcf\") " Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.521230 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wzmfj\" (UniqueName: \"kubernetes.io/projected/12cf87e0-5d1a-4598-bd14-e3ef92a56dc1-kube-api-access-wzmfj\") pod \"12cf87e0-5d1a-4598-bd14-e3ef92a56dc1\" (UID: \"12cf87e0-5d1a-4598-bd14-e3ef92a56dc1\") " Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.521256 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12cf87e0-5d1a-4598-bd14-e3ef92a56dc1-catalog-content\") pod \"12cf87e0-5d1a-4598-bd14-e3ef92a56dc1\" (UID: \"12cf87e0-5d1a-4598-bd14-e3ef92a56dc1\") " Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.521279 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b9e1de0-da0f-422e-a79b-4fb54d2e3acb-utilities\") pod \"1b9e1de0-da0f-422e-a79b-4fb54d2e3acb\" (UID: \"1b9e1de0-da0f-422e-a79b-4fb54d2e3acb\") " Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.521309 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65af6b35-41bd-4ad2-b8d4-0b653cb18dcf-catalog-content\") pod \"65af6b35-41bd-4ad2-b8d4-0b653cb18dcf\" (UID: \"65af6b35-41bd-4ad2-b8d4-0b653cb18dcf\") " Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.521339 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/74f92156-f803-433a-99a4-981202076e81-marketplace-operator-metrics\") pod \"74f92156-f803-433a-99a4-981202076e81\" (UID: \"74f92156-f803-433a-99a4-981202076e81\") " Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.521363 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b3a014a-3574-4c2b-8331-df326d47f240-utilities\") pod \"6b3a014a-3574-4c2b-8331-df326d47f240\" (UID: \"6b3a014a-3574-4c2b-8331-df326d47f240\") " Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.521420 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chhpl\" (UniqueName: \"kubernetes.io/projected/6b3a014a-3574-4c2b-8331-df326d47f240-kube-api-access-chhpl\") pod \"6b3a014a-3574-4c2b-8331-df326d47f240\" (UID: \"6b3a014a-3574-4c2b-8331-df326d47f240\") " Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.521456 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74f92156-f803-433a-99a4-981202076e81-marketplace-trusted-ca\") pod \"74f92156-f803-433a-99a4-981202076e81\" (UID: \"74f92156-f803-433a-99a4-981202076e81\") " Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.521482 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b9e1de0-da0f-422e-a79b-4fb54d2e3acb-catalog-content\") pod \"1b9e1de0-da0f-422e-a79b-4fb54d2e3acb\" (UID: \"1b9e1de0-da0f-422e-a79b-4fb54d2e3acb\") " Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.521525 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65af6b35-41bd-4ad2-b8d4-0b653cb18dcf-utilities\") pod \"65af6b35-41bd-4ad2-b8d4-0b653cb18dcf\" (UID: \"65af6b35-41bd-4ad2-b8d4-0b653cb18dcf\") " Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.524077 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12cf87e0-5d1a-4598-bd14-e3ef92a56dc1-kube-api-access-wzmfj" (OuterVolumeSpecName: "kube-api-access-wzmfj") pod "12cf87e0-5d1a-4598-bd14-e3ef92a56dc1" (UID: "12cf87e0-5d1a-4598-bd14-e3ef92a56dc1"). InnerVolumeSpecName "kube-api-access-wzmfj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.524303 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/74f92156-f803-433a-99a4-981202076e81-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "74f92156-f803-433a-99a4-981202076e81" (UID: "74f92156-f803-433a-99a4-981202076e81"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.524358 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74f92156-f803-433a-99a4-981202076e81-kube-api-access-q8hk6" (OuterVolumeSpecName: "kube-api-access-q8hk6") pod "74f92156-f803-433a-99a4-981202076e81" (UID: "74f92156-f803-433a-99a4-981202076e81"). InnerVolumeSpecName "kube-api-access-q8hk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.524707 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65af6b35-41bd-4ad2-b8d4-0b653cb18dcf-utilities" (OuterVolumeSpecName: "utilities") pod "65af6b35-41bd-4ad2-b8d4-0b653cb18dcf" (UID: "65af6b35-41bd-4ad2-b8d4-0b653cb18dcf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.524836 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b9e1de0-da0f-422e-a79b-4fb54d2e3acb-kube-api-access-tt5rs" (OuterVolumeSpecName: "kube-api-access-tt5rs") pod "1b9e1de0-da0f-422e-a79b-4fb54d2e3acb" (UID: "1b9e1de0-da0f-422e-a79b-4fb54d2e3acb"). InnerVolumeSpecName "kube-api-access-tt5rs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.525477 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12cf87e0-5d1a-4598-bd14-e3ef92a56dc1-utilities" (OuterVolumeSpecName: "utilities") pod "12cf87e0-5d1a-4598-bd14-e3ef92a56dc1" (UID: "12cf87e0-5d1a-4598-bd14-e3ef92a56dc1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.526100 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/65af6b35-41bd-4ad2-b8d4-0b653cb18dcf-kube-api-access-s6pmb" (OuterVolumeSpecName: "kube-api-access-s6pmb") pod "65af6b35-41bd-4ad2-b8d4-0b653cb18dcf" (UID: "65af6b35-41bd-4ad2-b8d4-0b653cb18dcf"). InnerVolumeSpecName "kube-api-access-s6pmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.526371 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74f92156-f803-433a-99a4-981202076e81-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "74f92156-f803-433a-99a4-981202076e81" (UID: "74f92156-f803-433a-99a4-981202076e81"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.529618 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b3a014a-3574-4c2b-8331-df326d47f240-kube-api-access-chhpl" (OuterVolumeSpecName: "kube-api-access-chhpl") pod "6b3a014a-3574-4c2b-8331-df326d47f240" (UID: "6b3a014a-3574-4c2b-8331-df326d47f240"). InnerVolumeSpecName "kube-api-access-chhpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.530219 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b3a014a-3574-4c2b-8331-df326d47f240-utilities" (OuterVolumeSpecName: "utilities") pod "6b3a014a-3574-4c2b-8331-df326d47f240" (UID: "6b3a014a-3574-4c2b-8331-df326d47f240"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.541617 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q8hk6\" (UniqueName: \"kubernetes.io/projected/74f92156-f803-433a-99a4-981202076e81-kube-api-access-q8hk6\") on node \"crc\" DevicePath \"\"" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.541657 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/12cf87e0-5d1a-4598-bd14-e3ef92a56dc1-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.541675 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tt5rs\" (UniqueName: \"kubernetes.io/projected/1b9e1de0-da0f-422e-a79b-4fb54d2e3acb-kube-api-access-tt5rs\") on node \"crc\" DevicePath \"\"" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.541687 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6pmb\" (UniqueName: \"kubernetes.io/projected/65af6b35-41bd-4ad2-b8d4-0b653cb18dcf-kube-api-access-s6pmb\") on node \"crc\" DevicePath \"\"" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.541698 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wzmfj\" (UniqueName: \"kubernetes.io/projected/12cf87e0-5d1a-4598-bd14-e3ef92a56dc1-kube-api-access-wzmfj\") on node \"crc\" DevicePath \"\"" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.541709 4813 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/74f92156-f803-433a-99a4-981202076e81-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.541720 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b3a014a-3574-4c2b-8331-df326d47f240-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.541731 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chhpl\" (UniqueName: \"kubernetes.io/projected/6b3a014a-3574-4c2b-8331-df326d47f240-kube-api-access-chhpl\") on node \"crc\" DevicePath \"\"" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.541745 4813 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/74f92156-f803-433a-99a4-981202076e81-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.541756 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/65af6b35-41bd-4ad2-b8d4-0b653cb18dcf-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.542766 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b9e1de0-da0f-422e-a79b-4fb54d2e3acb-utilities" (OuterVolumeSpecName: "utilities") pod "1b9e1de0-da0f-422e-a79b-4fb54d2e3acb" (UID: "1b9e1de0-da0f-422e-a79b-4fb54d2e3acb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.558740 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/12cf87e0-5d1a-4598-bd14-e3ef92a56dc1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "12cf87e0-5d1a-4598-bd14-e3ef92a56dc1" (UID: "12cf87e0-5d1a-4598-bd14-e3ef92a56dc1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.575066 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b9e1de0-da0f-422e-a79b-4fb54d2e3acb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b9e1de0-da0f-422e-a79b-4fb54d2e3acb" (UID: "1b9e1de0-da0f-422e-a79b-4fb54d2e3acb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.587406 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/65af6b35-41bd-4ad2-b8d4-0b653cb18dcf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "65af6b35-41bd-4ad2-b8d4-0b653cb18dcf" (UID: "65af6b35-41bd-4ad2-b8d4-0b653cb18dcf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.642792 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b9e1de0-da0f-422e-a79b-4fb54d2e3acb-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.643042 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/12cf87e0-5d1a-4598-bd14-e3ef92a56dc1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.643166 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b9e1de0-da0f-422e-a79b-4fb54d2e3acb-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.643255 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/65af6b35-41bd-4ad2-b8d4-0b653cb18dcf-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.683637 4813 generic.go:334] "Generic (PLEG): container finished" podID="74f92156-f803-433a-99a4-981202076e81" containerID="a2ff2e083beeae6c658a5db2db106c161d0759072bf21050e7653c63d2bf67d5" exitCode=0 Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.683785 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" event={"ID":"74f92156-f803-433a-99a4-981202076e81","Type":"ContainerDied","Data":"a2ff2e083beeae6c658a5db2db106c161d0759072bf21050e7653c63d2bf67d5"} Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.683843 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" event={"ID":"74f92156-f803-433a-99a4-981202076e81","Type":"ContainerDied","Data":"fca401c8c10481f6181cc4f34ed269a2838d6d313adba4fdfe6a786a51f43e33"} Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.683884 4813 scope.go:117] "RemoveContainer" containerID="a2ff2e083beeae6c658a5db2db106c161d0759072bf21050e7653c63d2bf67d5" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.684411 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5wl9g" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.686091 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b3a014a-3574-4c2b-8331-df326d47f240-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b3a014a-3574-4c2b-8331-df326d47f240" (UID: "6b3a014a-3574-4c2b-8331-df326d47f240"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.687399 4813 generic.go:334] "Generic (PLEG): container finished" podID="6b3a014a-3574-4c2b-8331-df326d47f240" containerID="3ccf2869040560c186883f68381aa6c02ceb23a455a78065f18335e2e2b78217" exitCode=0 Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.687517 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6gvsx" event={"ID":"6b3a014a-3574-4c2b-8331-df326d47f240","Type":"ContainerDied","Data":"3ccf2869040560c186883f68381aa6c02ceb23a455a78065f18335e2e2b78217"} Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.687555 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6gvsx" event={"ID":"6b3a014a-3574-4c2b-8331-df326d47f240","Type":"ContainerDied","Data":"e85159594d24618777672cced65c2e1452c4906e3c04bd52403f01727228e6fd"} Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.690001 4813 generic.go:334] "Generic (PLEG): container finished" podID="12cf87e0-5d1a-4598-bd14-e3ef92a56dc1" containerID="4e2f9e427afe63058f581dbffd89d35960f8c1034239db65888db4a7b6558688" exitCode=0 Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.690124 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-js422" event={"ID":"12cf87e0-5d1a-4598-bd14-e3ef92a56dc1","Type":"ContainerDied","Data":"4e2f9e427afe63058f581dbffd89d35960f8c1034239db65888db4a7b6558688"} Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.690145 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-js422" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.690169 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-js422" event={"ID":"12cf87e0-5d1a-4598-bd14-e3ef92a56dc1","Type":"ContainerDied","Data":"b3f529df3e9fb0355f893cedef76587326c48494900c6d1e92d79a85d5638f2d"} Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.690346 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6gvsx" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.692209 4813 generic.go:334] "Generic (PLEG): container finished" podID="1b9e1de0-da0f-422e-a79b-4fb54d2e3acb" containerID="76af5a3320222301f14aee13eb616c8cc0bfc67f5aca3d2f0701ca64340366b2" exitCode=0 Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.692336 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxrc7" event={"ID":"1b9e1de0-da0f-422e-a79b-4fb54d2e3acb","Type":"ContainerDied","Data":"76af5a3320222301f14aee13eb616c8cc0bfc67f5aca3d2f0701ca64340366b2"} Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.692379 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-lxrc7" event={"ID":"1b9e1de0-da0f-422e-a79b-4fb54d2e3acb","Type":"ContainerDied","Data":"8de6c471cf2943414655249287ef4a24d83e123ebcca15674b298e837c07d55d"} Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.692491 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-lxrc7" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.696129 4813 generic.go:334] "Generic (PLEG): container finished" podID="65af6b35-41bd-4ad2-b8d4-0b653cb18dcf" containerID="e61a91f1f051970c805f6541cb76afeeb6855918e43560a4f856e2e03dcf12d9" exitCode=0 Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.696173 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2skjk" event={"ID":"65af6b35-41bd-4ad2-b8d4-0b653cb18dcf","Type":"ContainerDied","Data":"e61a91f1f051970c805f6541cb76afeeb6855918e43560a4f856e2e03dcf12d9"} Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.696200 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2skjk" event={"ID":"65af6b35-41bd-4ad2-b8d4-0b653cb18dcf","Type":"ContainerDied","Data":"9d53eb7a90a322ac2991f4e21796bfce267b374b8627c60a7de116768b7f550d"} Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.696221 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2skjk" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.706724 4813 scope.go:117] "RemoveContainer" containerID="f6ab163fea4b95b91341d3e6ebaee0ae1e7c34b859dfe7cfc8c1d4b4c0e83595" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.728482 4813 scope.go:117] "RemoveContainer" containerID="a2ff2e083beeae6c658a5db2db106c161d0759072bf21050e7653c63d2bf67d5" Mar 17 09:17:31 crc kubenswrapper[4813]: E0317 09:17:31.730863 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2ff2e083beeae6c658a5db2db106c161d0759072bf21050e7653c63d2bf67d5\": container with ID starting with a2ff2e083beeae6c658a5db2db106c161d0759072bf21050e7653c63d2bf67d5 not found: ID does not exist" containerID="a2ff2e083beeae6c658a5db2db106c161d0759072bf21050e7653c63d2bf67d5" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.730902 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2ff2e083beeae6c658a5db2db106c161d0759072bf21050e7653c63d2bf67d5"} err="failed to get container status \"a2ff2e083beeae6c658a5db2db106c161d0759072bf21050e7653c63d2bf67d5\": rpc error: code = NotFound desc = could not find container \"a2ff2e083beeae6c658a5db2db106c161d0759072bf21050e7653c63d2bf67d5\": container with ID starting with a2ff2e083beeae6c658a5db2db106c161d0759072bf21050e7653c63d2bf67d5 not found: ID does not exist" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.730929 4813 scope.go:117] "RemoveContainer" containerID="f6ab163fea4b95b91341d3e6ebaee0ae1e7c34b859dfe7cfc8c1d4b4c0e83595" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.735795 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-nw44l"] Mar 17 09:17:31 crc kubenswrapper[4813]: E0317 09:17:31.736285 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6ab163fea4b95b91341d3e6ebaee0ae1e7c34b859dfe7cfc8c1d4b4c0e83595\": container with ID starting with f6ab163fea4b95b91341d3e6ebaee0ae1e7c34b859dfe7cfc8c1d4b4c0e83595 not found: ID does not exist" containerID="f6ab163fea4b95b91341d3e6ebaee0ae1e7c34b859dfe7cfc8c1d4b4c0e83595" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.736331 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6ab163fea4b95b91341d3e6ebaee0ae1e7c34b859dfe7cfc8c1d4b4c0e83595"} err="failed to get container status \"f6ab163fea4b95b91341d3e6ebaee0ae1e7c34b859dfe7cfc8c1d4b4c0e83595\": rpc error: code = NotFound desc = could not find container \"f6ab163fea4b95b91341d3e6ebaee0ae1e7c34b859dfe7cfc8c1d4b4c0e83595\": container with ID starting with f6ab163fea4b95b91341d3e6ebaee0ae1e7c34b859dfe7cfc8c1d4b4c0e83595 not found: ID does not exist" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.736362 4813 scope.go:117] "RemoveContainer" containerID="3ccf2869040560c186883f68381aa6c02ceb23a455a78065f18335e2e2b78217" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.744274 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b3a014a-3574-4c2b-8331-df326d47f240-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.746562 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5wl9g"] Mar 17 09:17:31 crc kubenswrapper[4813]: W0317 09:17:31.748057 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ca7f195_b5c8_4804_b95a_5ff53a637ac5.slice/crio-e3356ce60f3da45dbec47dfddf4eab3125a4697b6819802894d34879921dcff0 WatchSource:0}: Error finding container e3356ce60f3da45dbec47dfddf4eab3125a4697b6819802894d34879921dcff0: Status 404 returned error can't find the container with id e3356ce60f3da45dbec47dfddf4eab3125a4697b6819802894d34879921dcff0 Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.755267 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5wl9g"] Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.762914 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2skjk"] Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.768474 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2skjk"] Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.768520 4813 scope.go:117] "RemoveContainer" containerID="0f982775e6f3a4f0cea5af093812332c94e904833ea9524da65fc0bf13ce0552" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.779702 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6gvsx"] Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.788715 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6gvsx"] Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.802629 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-js422"] Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.813226 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-js422"] Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.817690 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-lxrc7"] Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.823878 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-lxrc7"] Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.824192 4813 scope.go:117] "RemoveContainer" containerID="d6d775b3a82cedca3c82e251809e2d700188f2a218b26d154e599e0ab0607c24" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.849581 4813 scope.go:117] "RemoveContainer" containerID="3ccf2869040560c186883f68381aa6c02ceb23a455a78065f18335e2e2b78217" Mar 17 09:17:31 crc kubenswrapper[4813]: E0317 09:17:31.850102 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ccf2869040560c186883f68381aa6c02ceb23a455a78065f18335e2e2b78217\": container with ID starting with 3ccf2869040560c186883f68381aa6c02ceb23a455a78065f18335e2e2b78217 not found: ID does not exist" containerID="3ccf2869040560c186883f68381aa6c02ceb23a455a78065f18335e2e2b78217" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.850170 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ccf2869040560c186883f68381aa6c02ceb23a455a78065f18335e2e2b78217"} err="failed to get container status \"3ccf2869040560c186883f68381aa6c02ceb23a455a78065f18335e2e2b78217\": rpc error: code = NotFound desc = could not find container \"3ccf2869040560c186883f68381aa6c02ceb23a455a78065f18335e2e2b78217\": container with ID starting with 3ccf2869040560c186883f68381aa6c02ceb23a455a78065f18335e2e2b78217 not found: ID does not exist" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.850314 4813 scope.go:117] "RemoveContainer" containerID="0f982775e6f3a4f0cea5af093812332c94e904833ea9524da65fc0bf13ce0552" Mar 17 09:17:31 crc kubenswrapper[4813]: E0317 09:17:31.850752 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0f982775e6f3a4f0cea5af093812332c94e904833ea9524da65fc0bf13ce0552\": container with ID starting with 0f982775e6f3a4f0cea5af093812332c94e904833ea9524da65fc0bf13ce0552 not found: ID does not exist" containerID="0f982775e6f3a4f0cea5af093812332c94e904833ea9524da65fc0bf13ce0552" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.850791 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0f982775e6f3a4f0cea5af093812332c94e904833ea9524da65fc0bf13ce0552"} err="failed to get container status \"0f982775e6f3a4f0cea5af093812332c94e904833ea9524da65fc0bf13ce0552\": rpc error: code = NotFound desc = could not find container \"0f982775e6f3a4f0cea5af093812332c94e904833ea9524da65fc0bf13ce0552\": container with ID starting with 0f982775e6f3a4f0cea5af093812332c94e904833ea9524da65fc0bf13ce0552 not found: ID does not exist" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.850819 4813 scope.go:117] "RemoveContainer" containerID="d6d775b3a82cedca3c82e251809e2d700188f2a218b26d154e599e0ab0607c24" Mar 17 09:17:31 crc kubenswrapper[4813]: E0317 09:17:31.851192 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d6d775b3a82cedca3c82e251809e2d700188f2a218b26d154e599e0ab0607c24\": container with ID starting with d6d775b3a82cedca3c82e251809e2d700188f2a218b26d154e599e0ab0607c24 not found: ID does not exist" containerID="d6d775b3a82cedca3c82e251809e2d700188f2a218b26d154e599e0ab0607c24" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.851245 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d6d775b3a82cedca3c82e251809e2d700188f2a218b26d154e599e0ab0607c24"} err="failed to get container status \"d6d775b3a82cedca3c82e251809e2d700188f2a218b26d154e599e0ab0607c24\": rpc error: code = NotFound desc = could not find container \"d6d775b3a82cedca3c82e251809e2d700188f2a218b26d154e599e0ab0607c24\": container with ID starting with d6d775b3a82cedca3c82e251809e2d700188f2a218b26d154e599e0ab0607c24 not found: ID does not exist" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.851267 4813 scope.go:117] "RemoveContainer" containerID="4e2f9e427afe63058f581dbffd89d35960f8c1034239db65888db4a7b6558688" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.881777 4813 scope.go:117] "RemoveContainer" containerID="1f30c4c96c804a7b9dc9a6b5d7a7558fc2822e74e875b763607ddb261b75d90b" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.904210 4813 scope.go:117] "RemoveContainer" containerID="20728e6f4ee982c6741a69a4a1b812c4e55d6d59c4b07372616ddaf75192f20b" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.920370 4813 scope.go:117] "RemoveContainer" containerID="4e2f9e427afe63058f581dbffd89d35960f8c1034239db65888db4a7b6558688" Mar 17 09:17:31 crc kubenswrapper[4813]: E0317 09:17:31.921007 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e2f9e427afe63058f581dbffd89d35960f8c1034239db65888db4a7b6558688\": container with ID starting with 4e2f9e427afe63058f581dbffd89d35960f8c1034239db65888db4a7b6558688 not found: ID does not exist" containerID="4e2f9e427afe63058f581dbffd89d35960f8c1034239db65888db4a7b6558688" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.921051 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e2f9e427afe63058f581dbffd89d35960f8c1034239db65888db4a7b6558688"} err="failed to get container status \"4e2f9e427afe63058f581dbffd89d35960f8c1034239db65888db4a7b6558688\": rpc error: code = NotFound desc = could not find container \"4e2f9e427afe63058f581dbffd89d35960f8c1034239db65888db4a7b6558688\": container with ID starting with 4e2f9e427afe63058f581dbffd89d35960f8c1034239db65888db4a7b6558688 not found: ID does not exist" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.921077 4813 scope.go:117] "RemoveContainer" containerID="1f30c4c96c804a7b9dc9a6b5d7a7558fc2822e74e875b763607ddb261b75d90b" Mar 17 09:17:31 crc kubenswrapper[4813]: E0317 09:17:31.921466 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f30c4c96c804a7b9dc9a6b5d7a7558fc2822e74e875b763607ddb261b75d90b\": container with ID starting with 1f30c4c96c804a7b9dc9a6b5d7a7558fc2822e74e875b763607ddb261b75d90b not found: ID does not exist" containerID="1f30c4c96c804a7b9dc9a6b5d7a7558fc2822e74e875b763607ddb261b75d90b" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.921488 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f30c4c96c804a7b9dc9a6b5d7a7558fc2822e74e875b763607ddb261b75d90b"} err="failed to get container status \"1f30c4c96c804a7b9dc9a6b5d7a7558fc2822e74e875b763607ddb261b75d90b\": rpc error: code = NotFound desc = could not find container \"1f30c4c96c804a7b9dc9a6b5d7a7558fc2822e74e875b763607ddb261b75d90b\": container with ID starting with 1f30c4c96c804a7b9dc9a6b5d7a7558fc2822e74e875b763607ddb261b75d90b not found: ID does not exist" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.921501 4813 scope.go:117] "RemoveContainer" containerID="20728e6f4ee982c6741a69a4a1b812c4e55d6d59c4b07372616ddaf75192f20b" Mar 17 09:17:31 crc kubenswrapper[4813]: E0317 09:17:31.922076 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20728e6f4ee982c6741a69a4a1b812c4e55d6d59c4b07372616ddaf75192f20b\": container with ID starting with 20728e6f4ee982c6741a69a4a1b812c4e55d6d59c4b07372616ddaf75192f20b not found: ID does not exist" containerID="20728e6f4ee982c6741a69a4a1b812c4e55d6d59c4b07372616ddaf75192f20b" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.922136 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20728e6f4ee982c6741a69a4a1b812c4e55d6d59c4b07372616ddaf75192f20b"} err="failed to get container status \"20728e6f4ee982c6741a69a4a1b812c4e55d6d59c4b07372616ddaf75192f20b\": rpc error: code = NotFound desc = could not find container \"20728e6f4ee982c6741a69a4a1b812c4e55d6d59c4b07372616ddaf75192f20b\": container with ID starting with 20728e6f4ee982c6741a69a4a1b812c4e55d6d59c4b07372616ddaf75192f20b not found: ID does not exist" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.922174 4813 scope.go:117] "RemoveContainer" containerID="76af5a3320222301f14aee13eb616c8cc0bfc67f5aca3d2f0701ca64340366b2" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.939136 4813 scope.go:117] "RemoveContainer" containerID="4d58dd63365d261f37b02064cbccd13904fbc1ce79697a65ff95c2fee7013b3e" Mar 17 09:17:31 crc kubenswrapper[4813]: I0317 09:17:31.957824 4813 scope.go:117] "RemoveContainer" containerID="22466df678fe5db67f486cbd971560fd45a8773a011b67e24d9fcbfe70ea5437" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.017817 4813 scope.go:117] "RemoveContainer" containerID="76af5a3320222301f14aee13eb616c8cc0bfc67f5aca3d2f0701ca64340366b2" Mar 17 09:17:32 crc kubenswrapper[4813]: E0317 09:17:32.018258 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76af5a3320222301f14aee13eb616c8cc0bfc67f5aca3d2f0701ca64340366b2\": container with ID starting with 76af5a3320222301f14aee13eb616c8cc0bfc67f5aca3d2f0701ca64340366b2 not found: ID does not exist" containerID="76af5a3320222301f14aee13eb616c8cc0bfc67f5aca3d2f0701ca64340366b2" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.018285 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76af5a3320222301f14aee13eb616c8cc0bfc67f5aca3d2f0701ca64340366b2"} err="failed to get container status \"76af5a3320222301f14aee13eb616c8cc0bfc67f5aca3d2f0701ca64340366b2\": rpc error: code = NotFound desc = could not find container \"76af5a3320222301f14aee13eb616c8cc0bfc67f5aca3d2f0701ca64340366b2\": container with ID starting with 76af5a3320222301f14aee13eb616c8cc0bfc67f5aca3d2f0701ca64340366b2 not found: ID does not exist" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.018305 4813 scope.go:117] "RemoveContainer" containerID="4d58dd63365d261f37b02064cbccd13904fbc1ce79697a65ff95c2fee7013b3e" Mar 17 09:17:32 crc kubenswrapper[4813]: E0317 09:17:32.018837 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d58dd63365d261f37b02064cbccd13904fbc1ce79697a65ff95c2fee7013b3e\": container with ID starting with 4d58dd63365d261f37b02064cbccd13904fbc1ce79697a65ff95c2fee7013b3e not found: ID does not exist" containerID="4d58dd63365d261f37b02064cbccd13904fbc1ce79697a65ff95c2fee7013b3e" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.018859 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d58dd63365d261f37b02064cbccd13904fbc1ce79697a65ff95c2fee7013b3e"} err="failed to get container status \"4d58dd63365d261f37b02064cbccd13904fbc1ce79697a65ff95c2fee7013b3e\": rpc error: code = NotFound desc = could not find container \"4d58dd63365d261f37b02064cbccd13904fbc1ce79697a65ff95c2fee7013b3e\": container with ID starting with 4d58dd63365d261f37b02064cbccd13904fbc1ce79697a65ff95c2fee7013b3e not found: ID does not exist" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.018876 4813 scope.go:117] "RemoveContainer" containerID="22466df678fe5db67f486cbd971560fd45a8773a011b67e24d9fcbfe70ea5437" Mar 17 09:17:32 crc kubenswrapper[4813]: E0317 09:17:32.019213 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"22466df678fe5db67f486cbd971560fd45a8773a011b67e24d9fcbfe70ea5437\": container with ID starting with 22466df678fe5db67f486cbd971560fd45a8773a011b67e24d9fcbfe70ea5437 not found: ID does not exist" containerID="22466df678fe5db67f486cbd971560fd45a8773a011b67e24d9fcbfe70ea5437" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.019236 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"22466df678fe5db67f486cbd971560fd45a8773a011b67e24d9fcbfe70ea5437"} err="failed to get container status \"22466df678fe5db67f486cbd971560fd45a8773a011b67e24d9fcbfe70ea5437\": rpc error: code = NotFound desc = could not find container \"22466df678fe5db67f486cbd971560fd45a8773a011b67e24d9fcbfe70ea5437\": container with ID starting with 22466df678fe5db67f486cbd971560fd45a8773a011b67e24d9fcbfe70ea5437 not found: ID does not exist" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.019254 4813 scope.go:117] "RemoveContainer" containerID="e61a91f1f051970c805f6541cb76afeeb6855918e43560a4f856e2e03dcf12d9" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.037577 4813 scope.go:117] "RemoveContainer" containerID="f10efea5d3325a7eeeeb9a544f307de5d03087d18da00fc581580837e1f39274" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.065453 4813 scope.go:117] "RemoveContainer" containerID="9711ceab7b56277dc92c943aff876676bcfc90a4859e2785dbaecb6b9be8644f" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.087742 4813 scope.go:117] "RemoveContainer" containerID="e61a91f1f051970c805f6541cb76afeeb6855918e43560a4f856e2e03dcf12d9" Mar 17 09:17:32 crc kubenswrapper[4813]: E0317 09:17:32.088132 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e61a91f1f051970c805f6541cb76afeeb6855918e43560a4f856e2e03dcf12d9\": container with ID starting with e61a91f1f051970c805f6541cb76afeeb6855918e43560a4f856e2e03dcf12d9 not found: ID does not exist" containerID="e61a91f1f051970c805f6541cb76afeeb6855918e43560a4f856e2e03dcf12d9" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.088162 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e61a91f1f051970c805f6541cb76afeeb6855918e43560a4f856e2e03dcf12d9"} err="failed to get container status \"e61a91f1f051970c805f6541cb76afeeb6855918e43560a4f856e2e03dcf12d9\": rpc error: code = NotFound desc = could not find container \"e61a91f1f051970c805f6541cb76afeeb6855918e43560a4f856e2e03dcf12d9\": container with ID starting with e61a91f1f051970c805f6541cb76afeeb6855918e43560a4f856e2e03dcf12d9 not found: ID does not exist" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.088185 4813 scope.go:117] "RemoveContainer" containerID="f10efea5d3325a7eeeeb9a544f307de5d03087d18da00fc581580837e1f39274" Mar 17 09:17:32 crc kubenswrapper[4813]: E0317 09:17:32.088536 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f10efea5d3325a7eeeeb9a544f307de5d03087d18da00fc581580837e1f39274\": container with ID starting with f10efea5d3325a7eeeeb9a544f307de5d03087d18da00fc581580837e1f39274 not found: ID does not exist" containerID="f10efea5d3325a7eeeeb9a544f307de5d03087d18da00fc581580837e1f39274" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.088555 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f10efea5d3325a7eeeeb9a544f307de5d03087d18da00fc581580837e1f39274"} err="failed to get container status \"f10efea5d3325a7eeeeb9a544f307de5d03087d18da00fc581580837e1f39274\": rpc error: code = NotFound desc = could not find container \"f10efea5d3325a7eeeeb9a544f307de5d03087d18da00fc581580837e1f39274\": container with ID starting with f10efea5d3325a7eeeeb9a544f307de5d03087d18da00fc581580837e1f39274 not found: ID does not exist" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.088568 4813 scope.go:117] "RemoveContainer" containerID="9711ceab7b56277dc92c943aff876676bcfc90a4859e2785dbaecb6b9be8644f" Mar 17 09:17:32 crc kubenswrapper[4813]: E0317 09:17:32.088956 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9711ceab7b56277dc92c943aff876676bcfc90a4859e2785dbaecb6b9be8644f\": container with ID starting with 9711ceab7b56277dc92c943aff876676bcfc90a4859e2785dbaecb6b9be8644f not found: ID does not exist" containerID="9711ceab7b56277dc92c943aff876676bcfc90a4859e2785dbaecb6b9be8644f" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.088983 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9711ceab7b56277dc92c943aff876676bcfc90a4859e2785dbaecb6b9be8644f"} err="failed to get container status \"9711ceab7b56277dc92c943aff876676bcfc90a4859e2785dbaecb6b9be8644f\": rpc error: code = NotFound desc = could not find container \"9711ceab7b56277dc92c943aff876676bcfc90a4859e2785dbaecb6b9be8644f\": container with ID starting with 9711ceab7b56277dc92c943aff876676bcfc90a4859e2785dbaecb6b9be8644f not found: ID does not exist" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.462314 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rshd8"] Mar 17 09:17:32 crc kubenswrapper[4813]: E0317 09:17:32.462508 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b3a014a-3574-4c2b-8331-df326d47f240" containerName="extract-content" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.462519 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b3a014a-3574-4c2b-8331-df326d47f240" containerName="extract-content" Mar 17 09:17:32 crc kubenswrapper[4813]: E0317 09:17:32.462530 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b3a014a-3574-4c2b-8331-df326d47f240" containerName="registry-server" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.462536 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b3a014a-3574-4c2b-8331-df326d47f240" containerName="registry-server" Mar 17 09:17:32 crc kubenswrapper[4813]: E0317 09:17:32.462546 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65af6b35-41bd-4ad2-b8d4-0b653cb18dcf" containerName="extract-content" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.462552 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="65af6b35-41bd-4ad2-b8d4-0b653cb18dcf" containerName="extract-content" Mar 17 09:17:32 crc kubenswrapper[4813]: E0317 09:17:32.462558 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74f92156-f803-433a-99a4-981202076e81" containerName="marketplace-operator" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.462564 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="74f92156-f803-433a-99a4-981202076e81" containerName="marketplace-operator" Mar 17 09:17:32 crc kubenswrapper[4813]: E0317 09:17:32.462570 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b9e1de0-da0f-422e-a79b-4fb54d2e3acb" containerName="extract-content" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.462587 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b9e1de0-da0f-422e-a79b-4fb54d2e3acb" containerName="extract-content" Mar 17 09:17:32 crc kubenswrapper[4813]: E0317 09:17:32.462606 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b3a014a-3574-4c2b-8331-df326d47f240" containerName="extract-utilities" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.462614 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b3a014a-3574-4c2b-8331-df326d47f240" containerName="extract-utilities" Mar 17 09:17:32 crc kubenswrapper[4813]: E0317 09:17:32.462623 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b9e1de0-da0f-422e-a79b-4fb54d2e3acb" containerName="extract-utilities" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.462628 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b9e1de0-da0f-422e-a79b-4fb54d2e3acb" containerName="extract-utilities" Mar 17 09:17:32 crc kubenswrapper[4813]: E0317 09:17:32.462638 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12cf87e0-5d1a-4598-bd14-e3ef92a56dc1" containerName="extract-content" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.462643 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="12cf87e0-5d1a-4598-bd14-e3ef92a56dc1" containerName="extract-content" Mar 17 09:17:32 crc kubenswrapper[4813]: E0317 09:17:32.462653 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12cf87e0-5d1a-4598-bd14-e3ef92a56dc1" containerName="registry-server" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.462659 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="12cf87e0-5d1a-4598-bd14-e3ef92a56dc1" containerName="registry-server" Mar 17 09:17:32 crc kubenswrapper[4813]: E0317 09:17:32.462668 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65af6b35-41bd-4ad2-b8d4-0b653cb18dcf" containerName="extract-utilities" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.462673 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="65af6b35-41bd-4ad2-b8d4-0b653cb18dcf" containerName="extract-utilities" Mar 17 09:17:32 crc kubenswrapper[4813]: E0317 09:17:32.462684 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12cf87e0-5d1a-4598-bd14-e3ef92a56dc1" containerName="extract-utilities" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.462689 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="12cf87e0-5d1a-4598-bd14-e3ef92a56dc1" containerName="extract-utilities" Mar 17 09:17:32 crc kubenswrapper[4813]: E0317 09:17:32.462699 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b9e1de0-da0f-422e-a79b-4fb54d2e3acb" containerName="registry-server" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.462704 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b9e1de0-da0f-422e-a79b-4fb54d2e3acb" containerName="registry-server" Mar 17 09:17:32 crc kubenswrapper[4813]: E0317 09:17:32.462710 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="65af6b35-41bd-4ad2-b8d4-0b653cb18dcf" containerName="registry-server" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.462715 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="65af6b35-41bd-4ad2-b8d4-0b653cb18dcf" containerName="registry-server" Mar 17 09:17:32 crc kubenswrapper[4813]: E0317 09:17:32.462723 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74f92156-f803-433a-99a4-981202076e81" containerName="marketplace-operator" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.462729 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="74f92156-f803-433a-99a4-981202076e81" containerName="marketplace-operator" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.462803 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="74f92156-f803-433a-99a4-981202076e81" containerName="marketplace-operator" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.462814 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b3a014a-3574-4c2b-8331-df326d47f240" containerName="registry-server" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.462821 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="74f92156-f803-433a-99a4-981202076e81" containerName="marketplace-operator" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.462828 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="12cf87e0-5d1a-4598-bd14-e3ef92a56dc1" containerName="registry-server" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.462837 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="65af6b35-41bd-4ad2-b8d4-0b653cb18dcf" containerName="registry-server" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.462848 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b9e1de0-da0f-422e-a79b-4fb54d2e3acb" containerName="registry-server" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.463537 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rshd8" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.465502 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.475513 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rshd8"] Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.654180 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a972da79-2f83-434c-8466-d59b21923a84-utilities\") pod \"community-operators-rshd8\" (UID: \"a972da79-2f83-434c-8466-d59b21923a84\") " pod="openshift-marketplace/community-operators-rshd8" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.654263 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8rzv2\" (UniqueName: \"kubernetes.io/projected/a972da79-2f83-434c-8466-d59b21923a84-kube-api-access-8rzv2\") pod \"community-operators-rshd8\" (UID: \"a972da79-2f83-434c-8466-d59b21923a84\") " pod="openshift-marketplace/community-operators-rshd8" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.654317 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a972da79-2f83-434c-8466-d59b21923a84-catalog-content\") pod \"community-operators-rshd8\" (UID: \"a972da79-2f83-434c-8466-d59b21923a84\") " pod="openshift-marketplace/community-operators-rshd8" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.721325 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nw44l" event={"ID":"6ca7f195-b5c8-4804-b95a-5ff53a637ac5","Type":"ContainerStarted","Data":"cb52420c281af1b2ff601ace178211d8bd5e3e67d1db0fc21b413faff83f07f7"} Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.721383 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-nw44l" event={"ID":"6ca7f195-b5c8-4804-b95a-5ff53a637ac5","Type":"ContainerStarted","Data":"e3356ce60f3da45dbec47dfddf4eab3125a4697b6819802894d34879921dcff0"} Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.725423 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-nw44l" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.749836 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12cf87e0-5d1a-4598-bd14-e3ef92a56dc1" path="/var/lib/kubelet/pods/12cf87e0-5d1a-4598-bd14-e3ef92a56dc1/volumes" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.752083 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b9e1de0-da0f-422e-a79b-4fb54d2e3acb" path="/var/lib/kubelet/pods/1b9e1de0-da0f-422e-a79b-4fb54d2e3acb/volumes" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.754498 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="65af6b35-41bd-4ad2-b8d4-0b653cb18dcf" path="/var/lib/kubelet/pods/65af6b35-41bd-4ad2-b8d4-0b653cb18dcf/volumes" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.756281 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a972da79-2f83-434c-8466-d59b21923a84-catalog-content\") pod \"community-operators-rshd8\" (UID: \"a972da79-2f83-434c-8466-d59b21923a84\") " pod="openshift-marketplace/community-operators-rshd8" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.756624 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a972da79-2f83-434c-8466-d59b21923a84-utilities\") pod \"community-operators-rshd8\" (UID: \"a972da79-2f83-434c-8466-d59b21923a84\") " pod="openshift-marketplace/community-operators-rshd8" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.757199 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a972da79-2f83-434c-8466-d59b21923a84-utilities\") pod \"community-operators-rshd8\" (UID: \"a972da79-2f83-434c-8466-d59b21923a84\") " pod="openshift-marketplace/community-operators-rshd8" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.757218 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8rzv2\" (UniqueName: \"kubernetes.io/projected/a972da79-2f83-434c-8466-d59b21923a84-kube-api-access-8rzv2\") pod \"community-operators-rshd8\" (UID: \"a972da79-2f83-434c-8466-d59b21923a84\") " pod="openshift-marketplace/community-operators-rshd8" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.757295 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a972da79-2f83-434c-8466-d59b21923a84-catalog-content\") pod \"community-operators-rshd8\" (UID: \"a972da79-2f83-434c-8466-d59b21923a84\") " pod="openshift-marketplace/community-operators-rshd8" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.758259 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b3a014a-3574-4c2b-8331-df326d47f240" path="/var/lib/kubelet/pods/6b3a014a-3574-4c2b-8331-df326d47f240/volumes" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.759502 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74f92156-f803-433a-99a4-981202076e81" path="/var/lib/kubelet/pods/74f92156-f803-433a-99a4-981202076e81/volumes" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.760907 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-nw44l" podStartSLOduration=2.760845554 podStartE2EDuration="2.760845554s" podCreationTimestamp="2026-03-17 09:17:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:17:32.750528375 +0000 UTC m=+474.851331914" watchObservedRunningTime="2026-03-17 09:17:32.760845554 +0000 UTC m=+474.861649083" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.761140 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-nw44l" Mar 17 09:17:32 crc kubenswrapper[4813]: I0317 09:17:32.783867 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8rzv2\" (UniqueName: \"kubernetes.io/projected/a972da79-2f83-434c-8466-d59b21923a84-kube-api-access-8rzv2\") pod \"community-operators-rshd8\" (UID: \"a972da79-2f83-434c-8466-d59b21923a84\") " pod="openshift-marketplace/community-operators-rshd8" Mar 17 09:17:33 crc kubenswrapper[4813]: I0317 09:17:33.083117 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rshd8" Mar 17 09:17:33 crc kubenswrapper[4813]: I0317 09:17:33.319314 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rshd8"] Mar 17 09:17:33 crc kubenswrapper[4813]: W0317 09:17:33.322213 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda972da79_2f83_434c_8466_d59b21923a84.slice/crio-dfcc7321fb92c0213f31c83e00833698ba1f889f3dfb35063062181c9673922e WatchSource:0}: Error finding container dfcc7321fb92c0213f31c83e00833698ba1f889f3dfb35063062181c9673922e: Status 404 returned error can't find the container with id dfcc7321fb92c0213f31c83e00833698ba1f889f3dfb35063062181c9673922e Mar 17 09:17:33 crc kubenswrapper[4813]: I0317 09:17:33.729305 4813 generic.go:334] "Generic (PLEG): container finished" podID="a972da79-2f83-434c-8466-d59b21923a84" containerID="821d50bf0926f340da658e2ab21b2189e3de607da0f6c3df7209cd6222ec2db4" exitCode=0 Mar 17 09:17:33 crc kubenswrapper[4813]: I0317 09:17:33.729341 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rshd8" event={"ID":"a972da79-2f83-434c-8466-d59b21923a84","Type":"ContainerDied","Data":"821d50bf0926f340da658e2ab21b2189e3de607da0f6c3df7209cd6222ec2db4"} Mar 17 09:17:33 crc kubenswrapper[4813]: I0317 09:17:33.729586 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rshd8" event={"ID":"a972da79-2f83-434c-8466-d59b21923a84","Type":"ContainerStarted","Data":"dfcc7321fb92c0213f31c83e00833698ba1f889f3dfb35063062181c9673922e"} Mar 17 09:17:33 crc kubenswrapper[4813]: I0317 09:17:33.872557 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-j7m85"] Mar 17 09:17:33 crc kubenswrapper[4813]: I0317 09:17:33.876505 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j7m85" Mar 17 09:17:33 crc kubenswrapper[4813]: I0317 09:17:33.881099 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 17 09:17:33 crc kubenswrapper[4813]: I0317 09:17:33.881430 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j7m85"] Mar 17 09:17:33 crc kubenswrapper[4813]: I0317 09:17:33.899015 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7hkx\" (UniqueName: \"kubernetes.io/projected/a15c3d28-d8aa-4d38-a180-420c05345960-kube-api-access-h7hkx\") pod \"certified-operators-j7m85\" (UID: \"a15c3d28-d8aa-4d38-a180-420c05345960\") " pod="openshift-marketplace/certified-operators-j7m85" Mar 17 09:17:33 crc kubenswrapper[4813]: I0317 09:17:33.899085 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a15c3d28-d8aa-4d38-a180-420c05345960-catalog-content\") pod \"certified-operators-j7m85\" (UID: \"a15c3d28-d8aa-4d38-a180-420c05345960\") " pod="openshift-marketplace/certified-operators-j7m85" Mar 17 09:17:33 crc kubenswrapper[4813]: I0317 09:17:33.899117 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a15c3d28-d8aa-4d38-a180-420c05345960-utilities\") pod \"certified-operators-j7m85\" (UID: \"a15c3d28-d8aa-4d38-a180-420c05345960\") " pod="openshift-marketplace/certified-operators-j7m85" Mar 17 09:17:34 crc kubenswrapper[4813]: I0317 09:17:34.000413 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a15c3d28-d8aa-4d38-a180-420c05345960-catalog-content\") pod \"certified-operators-j7m85\" (UID: \"a15c3d28-d8aa-4d38-a180-420c05345960\") " pod="openshift-marketplace/certified-operators-j7m85" Mar 17 09:17:34 crc kubenswrapper[4813]: I0317 09:17:34.000464 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a15c3d28-d8aa-4d38-a180-420c05345960-utilities\") pod \"certified-operators-j7m85\" (UID: \"a15c3d28-d8aa-4d38-a180-420c05345960\") " pod="openshift-marketplace/certified-operators-j7m85" Mar 17 09:17:34 crc kubenswrapper[4813]: I0317 09:17:34.000549 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7hkx\" (UniqueName: \"kubernetes.io/projected/a15c3d28-d8aa-4d38-a180-420c05345960-kube-api-access-h7hkx\") pod \"certified-operators-j7m85\" (UID: \"a15c3d28-d8aa-4d38-a180-420c05345960\") " pod="openshift-marketplace/certified-operators-j7m85" Mar 17 09:17:34 crc kubenswrapper[4813]: I0317 09:17:34.000981 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a15c3d28-d8aa-4d38-a180-420c05345960-catalog-content\") pod \"certified-operators-j7m85\" (UID: \"a15c3d28-d8aa-4d38-a180-420c05345960\") " pod="openshift-marketplace/certified-operators-j7m85" Mar 17 09:17:34 crc kubenswrapper[4813]: I0317 09:17:34.001009 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a15c3d28-d8aa-4d38-a180-420c05345960-utilities\") pod \"certified-operators-j7m85\" (UID: \"a15c3d28-d8aa-4d38-a180-420c05345960\") " pod="openshift-marketplace/certified-operators-j7m85" Mar 17 09:17:34 crc kubenswrapper[4813]: I0317 09:17:34.020153 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7hkx\" (UniqueName: \"kubernetes.io/projected/a15c3d28-d8aa-4d38-a180-420c05345960-kube-api-access-h7hkx\") pod \"certified-operators-j7m85\" (UID: \"a15c3d28-d8aa-4d38-a180-420c05345960\") " pod="openshift-marketplace/certified-operators-j7m85" Mar 17 09:17:34 crc kubenswrapper[4813]: I0317 09:17:34.212267 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j7m85" Mar 17 09:17:34 crc kubenswrapper[4813]: I0317 09:17:34.652615 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-j7m85"] Mar 17 09:17:34 crc kubenswrapper[4813]: W0317 09:17:34.658678 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda15c3d28_d8aa_4d38_a180_420c05345960.slice/crio-235123a2e4e9b20c943afb86d8c6d8956d0d581a86d85bbcfe7615b3b049e46a WatchSource:0}: Error finding container 235123a2e4e9b20c943afb86d8c6d8956d0d581a86d85bbcfe7615b3b049e46a: Status 404 returned error can't find the container with id 235123a2e4e9b20c943afb86d8c6d8956d0d581a86d85bbcfe7615b3b049e46a Mar 17 09:17:34 crc kubenswrapper[4813]: I0317 09:17:34.740502 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7m85" event={"ID":"a15c3d28-d8aa-4d38-a180-420c05345960","Type":"ContainerStarted","Data":"235123a2e4e9b20c943afb86d8c6d8956d0d581a86d85bbcfe7615b3b049e46a"} Mar 17 09:17:34 crc kubenswrapper[4813]: I0317 09:17:34.866183 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-54j2x"] Mar 17 09:17:34 crc kubenswrapper[4813]: I0317 09:17:34.867701 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-54j2x" Mar 17 09:17:34 crc kubenswrapper[4813]: I0317 09:17:34.871426 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 17 09:17:34 crc kubenswrapper[4813]: I0317 09:17:34.877541 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-54j2x"] Mar 17 09:17:34 crc kubenswrapper[4813]: I0317 09:17:34.915972 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7e12568-6cf2-4551-8aff-b330c8eda201-utilities\") pod \"redhat-marketplace-54j2x\" (UID: \"c7e12568-6cf2-4551-8aff-b330c8eda201\") " pod="openshift-marketplace/redhat-marketplace-54j2x" Mar 17 09:17:34 crc kubenswrapper[4813]: I0317 09:17:34.916317 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7e12568-6cf2-4551-8aff-b330c8eda201-catalog-content\") pod \"redhat-marketplace-54j2x\" (UID: \"c7e12568-6cf2-4551-8aff-b330c8eda201\") " pod="openshift-marketplace/redhat-marketplace-54j2x" Mar 17 09:17:34 crc kubenswrapper[4813]: I0317 09:17:34.916526 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qh98s\" (UniqueName: \"kubernetes.io/projected/c7e12568-6cf2-4551-8aff-b330c8eda201-kube-api-access-qh98s\") pod \"redhat-marketplace-54j2x\" (UID: \"c7e12568-6cf2-4551-8aff-b330c8eda201\") " pod="openshift-marketplace/redhat-marketplace-54j2x" Mar 17 09:17:35 crc kubenswrapper[4813]: I0317 09:17:35.017725 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7e12568-6cf2-4551-8aff-b330c8eda201-catalog-content\") pod \"redhat-marketplace-54j2x\" (UID: \"c7e12568-6cf2-4551-8aff-b330c8eda201\") " pod="openshift-marketplace/redhat-marketplace-54j2x" Mar 17 09:17:35 crc kubenswrapper[4813]: I0317 09:17:35.017792 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qh98s\" (UniqueName: \"kubernetes.io/projected/c7e12568-6cf2-4551-8aff-b330c8eda201-kube-api-access-qh98s\") pod \"redhat-marketplace-54j2x\" (UID: \"c7e12568-6cf2-4551-8aff-b330c8eda201\") " pod="openshift-marketplace/redhat-marketplace-54j2x" Mar 17 09:17:35 crc kubenswrapper[4813]: I0317 09:17:35.017853 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7e12568-6cf2-4551-8aff-b330c8eda201-utilities\") pod \"redhat-marketplace-54j2x\" (UID: \"c7e12568-6cf2-4551-8aff-b330c8eda201\") " pod="openshift-marketplace/redhat-marketplace-54j2x" Mar 17 09:17:35 crc kubenswrapper[4813]: I0317 09:17:35.018194 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c7e12568-6cf2-4551-8aff-b330c8eda201-catalog-content\") pod \"redhat-marketplace-54j2x\" (UID: \"c7e12568-6cf2-4551-8aff-b330c8eda201\") " pod="openshift-marketplace/redhat-marketplace-54j2x" Mar 17 09:17:35 crc kubenswrapper[4813]: I0317 09:17:35.018236 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c7e12568-6cf2-4551-8aff-b330c8eda201-utilities\") pod \"redhat-marketplace-54j2x\" (UID: \"c7e12568-6cf2-4551-8aff-b330c8eda201\") " pod="openshift-marketplace/redhat-marketplace-54j2x" Mar 17 09:17:35 crc kubenswrapper[4813]: I0317 09:17:35.040917 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qh98s\" (UniqueName: \"kubernetes.io/projected/c7e12568-6cf2-4551-8aff-b330c8eda201-kube-api-access-qh98s\") pod \"redhat-marketplace-54j2x\" (UID: \"c7e12568-6cf2-4551-8aff-b330c8eda201\") " pod="openshift-marketplace/redhat-marketplace-54j2x" Mar 17 09:17:35 crc kubenswrapper[4813]: I0317 09:17:35.240263 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-54j2x" Mar 17 09:17:35 crc kubenswrapper[4813]: I0317 09:17:35.503738 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-54j2x"] Mar 17 09:17:35 crc kubenswrapper[4813]: W0317 09:17:35.512020 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc7e12568_6cf2_4551_8aff_b330c8eda201.slice/crio-a12a09e159128bc3464e116c41d05f4140e93ae113d87ed6e7b8b7450535ec79 WatchSource:0}: Error finding container a12a09e159128bc3464e116c41d05f4140e93ae113d87ed6e7b8b7450535ec79: Status 404 returned error can't find the container with id a12a09e159128bc3464e116c41d05f4140e93ae113d87ed6e7b8b7450535ec79 Mar 17 09:17:35 crc kubenswrapper[4813]: I0317 09:17:35.748845 4813 generic.go:334] "Generic (PLEG): container finished" podID="a972da79-2f83-434c-8466-d59b21923a84" containerID="c9bd84b9903ae3f96730a1048c03f6d9da637aaf6cab6649578e9e4bc4f99d18" exitCode=0 Mar 17 09:17:35 crc kubenswrapper[4813]: I0317 09:17:35.748911 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rshd8" event={"ID":"a972da79-2f83-434c-8466-d59b21923a84","Type":"ContainerDied","Data":"c9bd84b9903ae3f96730a1048c03f6d9da637aaf6cab6649578e9e4bc4f99d18"} Mar 17 09:17:35 crc kubenswrapper[4813]: I0317 09:17:35.754390 4813 generic.go:334] "Generic (PLEG): container finished" podID="a15c3d28-d8aa-4d38-a180-420c05345960" containerID="5e5caf98ec1b4ebbec66e61c1ee07e96b7274f80315ba8aed847348d5c9c8109" exitCode=0 Mar 17 09:17:35 crc kubenswrapper[4813]: I0317 09:17:35.754498 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7m85" event={"ID":"a15c3d28-d8aa-4d38-a180-420c05345960","Type":"ContainerDied","Data":"5e5caf98ec1b4ebbec66e61c1ee07e96b7274f80315ba8aed847348d5c9c8109"} Mar 17 09:17:35 crc kubenswrapper[4813]: I0317 09:17:35.757690 4813 generic.go:334] "Generic (PLEG): container finished" podID="c7e12568-6cf2-4551-8aff-b330c8eda201" containerID="f076f34b00ac00727f3fe1a655629c9ed4095d7da4e31225bb51cc256049476c" exitCode=0 Mar 17 09:17:35 crc kubenswrapper[4813]: I0317 09:17:35.757721 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54j2x" event={"ID":"c7e12568-6cf2-4551-8aff-b330c8eda201","Type":"ContainerDied","Data":"f076f34b00ac00727f3fe1a655629c9ed4095d7da4e31225bb51cc256049476c"} Mar 17 09:17:35 crc kubenswrapper[4813]: I0317 09:17:35.759075 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54j2x" event={"ID":"c7e12568-6cf2-4551-8aff-b330c8eda201","Type":"ContainerStarted","Data":"a12a09e159128bc3464e116c41d05f4140e93ae113d87ed6e7b8b7450535ec79"} Mar 17 09:17:36 crc kubenswrapper[4813]: I0317 09:17:36.272783 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-cm7vw"] Mar 17 09:17:36 crc kubenswrapper[4813]: I0317 09:17:36.291476 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cm7vw" Mar 17 09:17:36 crc kubenswrapper[4813]: I0317 09:17:36.299123 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 17 09:17:36 crc kubenswrapper[4813]: I0317 09:17:36.309189 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cm7vw"] Mar 17 09:17:36 crc kubenswrapper[4813]: I0317 09:17:36.435752 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sffwb\" (UniqueName: \"kubernetes.io/projected/ba5c8bcc-1783-4889-a504-4233deca14dd-kube-api-access-sffwb\") pod \"redhat-operators-cm7vw\" (UID: \"ba5c8bcc-1783-4889-a504-4233deca14dd\") " pod="openshift-marketplace/redhat-operators-cm7vw" Mar 17 09:17:36 crc kubenswrapper[4813]: I0317 09:17:36.435807 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba5c8bcc-1783-4889-a504-4233deca14dd-catalog-content\") pod \"redhat-operators-cm7vw\" (UID: \"ba5c8bcc-1783-4889-a504-4233deca14dd\") " pod="openshift-marketplace/redhat-operators-cm7vw" Mar 17 09:17:36 crc kubenswrapper[4813]: I0317 09:17:36.435859 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba5c8bcc-1783-4889-a504-4233deca14dd-utilities\") pod \"redhat-operators-cm7vw\" (UID: \"ba5c8bcc-1783-4889-a504-4233deca14dd\") " pod="openshift-marketplace/redhat-operators-cm7vw" Mar 17 09:17:36 crc kubenswrapper[4813]: I0317 09:17:36.537309 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba5c8bcc-1783-4889-a504-4233deca14dd-utilities\") pod \"redhat-operators-cm7vw\" (UID: \"ba5c8bcc-1783-4889-a504-4233deca14dd\") " pod="openshift-marketplace/redhat-operators-cm7vw" Mar 17 09:17:36 crc kubenswrapper[4813]: I0317 09:17:36.537388 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sffwb\" (UniqueName: \"kubernetes.io/projected/ba5c8bcc-1783-4889-a504-4233deca14dd-kube-api-access-sffwb\") pod \"redhat-operators-cm7vw\" (UID: \"ba5c8bcc-1783-4889-a504-4233deca14dd\") " pod="openshift-marketplace/redhat-operators-cm7vw" Mar 17 09:17:36 crc kubenswrapper[4813]: I0317 09:17:36.537423 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba5c8bcc-1783-4889-a504-4233deca14dd-catalog-content\") pod \"redhat-operators-cm7vw\" (UID: \"ba5c8bcc-1783-4889-a504-4233deca14dd\") " pod="openshift-marketplace/redhat-operators-cm7vw" Mar 17 09:17:36 crc kubenswrapper[4813]: I0317 09:17:36.537875 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba5c8bcc-1783-4889-a504-4233deca14dd-catalog-content\") pod \"redhat-operators-cm7vw\" (UID: \"ba5c8bcc-1783-4889-a504-4233deca14dd\") " pod="openshift-marketplace/redhat-operators-cm7vw" Mar 17 09:17:36 crc kubenswrapper[4813]: I0317 09:17:36.538341 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba5c8bcc-1783-4889-a504-4233deca14dd-utilities\") pod \"redhat-operators-cm7vw\" (UID: \"ba5c8bcc-1783-4889-a504-4233deca14dd\") " pod="openshift-marketplace/redhat-operators-cm7vw" Mar 17 09:17:36 crc kubenswrapper[4813]: I0317 09:17:36.565738 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sffwb\" (UniqueName: \"kubernetes.io/projected/ba5c8bcc-1783-4889-a504-4233deca14dd-kube-api-access-sffwb\") pod \"redhat-operators-cm7vw\" (UID: \"ba5c8bcc-1783-4889-a504-4233deca14dd\") " pod="openshift-marketplace/redhat-operators-cm7vw" Mar 17 09:17:36 crc kubenswrapper[4813]: I0317 09:17:36.617868 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cm7vw" Mar 17 09:17:36 crc kubenswrapper[4813]: I0317 09:17:36.619341 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-m59rk" Mar 17 09:17:36 crc kubenswrapper[4813]: I0317 09:17:36.678490 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dbm4b"] Mar 17 09:17:36 crc kubenswrapper[4813]: I0317 09:17:36.784639 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54j2x" event={"ID":"c7e12568-6cf2-4551-8aff-b330c8eda201","Type":"ContainerStarted","Data":"5b9f80ac7c8a727724818f2588271f2e832b0dc54cbcd704dee949e0d2aa140e"} Mar 17 09:17:36 crc kubenswrapper[4813]: I0317 09:17:36.789746 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rshd8" event={"ID":"a972da79-2f83-434c-8466-d59b21923a84","Type":"ContainerStarted","Data":"83013ad1063315714c2b36d68b2f80061a8950bdfa29ed7fd9c4c370263e506f"} Mar 17 09:17:36 crc kubenswrapper[4813]: I0317 09:17:36.796506 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7m85" event={"ID":"a15c3d28-d8aa-4d38-a180-420c05345960","Type":"ContainerStarted","Data":"0d4083d119d8c44805c93913a7f730a5c5bd90393faa5d3a37a1476d8d237c01"} Mar 17 09:17:36 crc kubenswrapper[4813]: I0317 09:17:36.841485 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rshd8" podStartSLOduration=2.088648949 podStartE2EDuration="4.841463412s" podCreationTimestamp="2026-03-17 09:17:32 +0000 UTC" firstStartedPulling="2026-03-17 09:17:33.732004256 +0000 UTC m=+475.832807755" lastFinishedPulling="2026-03-17 09:17:36.484818719 +0000 UTC m=+478.585622218" observedRunningTime="2026-03-17 09:17:36.839165041 +0000 UTC m=+478.939968540" watchObservedRunningTime="2026-03-17 09:17:36.841463412 +0000 UTC m=+478.942266911" Mar 17 09:17:37 crc kubenswrapper[4813]: I0317 09:17:36.933877 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-cm7vw"] Mar 17 09:17:37 crc kubenswrapper[4813]: I0317 09:17:37.804940 4813 generic.go:334] "Generic (PLEG): container finished" podID="ba5c8bcc-1783-4889-a504-4233deca14dd" containerID="4acaa380639f15fa272d0ea3f564e8fa0e2a137daa7f0b4d711c533350647f7c" exitCode=0 Mar 17 09:17:37 crc kubenswrapper[4813]: I0317 09:17:37.805029 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cm7vw" event={"ID":"ba5c8bcc-1783-4889-a504-4233deca14dd","Type":"ContainerDied","Data":"4acaa380639f15fa272d0ea3f564e8fa0e2a137daa7f0b4d711c533350647f7c"} Mar 17 09:17:37 crc kubenswrapper[4813]: I0317 09:17:37.805067 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cm7vw" event={"ID":"ba5c8bcc-1783-4889-a504-4233deca14dd","Type":"ContainerStarted","Data":"cec77bbf462499140275b2ada28ce2e75b9abd072218f04f19842ae47a50f941"} Mar 17 09:17:37 crc kubenswrapper[4813]: I0317 09:17:37.807207 4813 generic.go:334] "Generic (PLEG): container finished" podID="a15c3d28-d8aa-4d38-a180-420c05345960" containerID="0d4083d119d8c44805c93913a7f730a5c5bd90393faa5d3a37a1476d8d237c01" exitCode=0 Mar 17 09:17:37 crc kubenswrapper[4813]: I0317 09:17:37.807260 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7m85" event={"ID":"a15c3d28-d8aa-4d38-a180-420c05345960","Type":"ContainerDied","Data":"0d4083d119d8c44805c93913a7f730a5c5bd90393faa5d3a37a1476d8d237c01"} Mar 17 09:17:37 crc kubenswrapper[4813]: I0317 09:17:37.810066 4813 generic.go:334] "Generic (PLEG): container finished" podID="c7e12568-6cf2-4551-8aff-b330c8eda201" containerID="5b9f80ac7c8a727724818f2588271f2e832b0dc54cbcd704dee949e0d2aa140e" exitCode=0 Mar 17 09:17:37 crc kubenswrapper[4813]: I0317 09:17:37.810138 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54j2x" event={"ID":"c7e12568-6cf2-4551-8aff-b330c8eda201","Type":"ContainerDied","Data":"5b9f80ac7c8a727724818f2588271f2e832b0dc54cbcd704dee949e0d2aa140e"} Mar 17 09:17:38 crc kubenswrapper[4813]: I0317 09:17:38.822833 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54j2x" event={"ID":"c7e12568-6cf2-4551-8aff-b330c8eda201","Type":"ContainerStarted","Data":"41d4f1ac125feeff978fa0c68d90daf196f35d1b41500de22e346743fb25b109"} Mar 17 09:17:39 crc kubenswrapper[4813]: I0317 09:17:39.829889 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7m85" event={"ID":"a15c3d28-d8aa-4d38-a180-420c05345960","Type":"ContainerStarted","Data":"38de3e851b2a82e6206e5adf6ce5348414401b6d3e6657ac5ef2249d45f2a213"} Mar 17 09:17:39 crc kubenswrapper[4813]: I0317 09:17:39.832191 4813 generic.go:334] "Generic (PLEG): container finished" podID="ba5c8bcc-1783-4889-a504-4233deca14dd" containerID="bc81c19ce8a9d9e4f3ab3c76d4baf2d0f737dc5997af0720ec4b525de2db7ccf" exitCode=0 Mar 17 09:17:39 crc kubenswrapper[4813]: I0317 09:17:39.833378 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cm7vw" event={"ID":"ba5c8bcc-1783-4889-a504-4233deca14dd","Type":"ContainerDied","Data":"bc81c19ce8a9d9e4f3ab3c76d4baf2d0f737dc5997af0720ec4b525de2db7ccf"} Mar 17 09:17:39 crc kubenswrapper[4813]: I0317 09:17:39.853506 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-j7m85" podStartSLOduration=3.960410325 podStartE2EDuration="6.853487409s" podCreationTimestamp="2026-03-17 09:17:33 +0000 UTC" firstStartedPulling="2026-03-17 09:17:35.756500596 +0000 UTC m=+477.857304105" lastFinishedPulling="2026-03-17 09:17:38.64957769 +0000 UTC m=+480.750381189" observedRunningTime="2026-03-17 09:17:39.852296983 +0000 UTC m=+481.953100492" watchObservedRunningTime="2026-03-17 09:17:39.853487409 +0000 UTC m=+481.954290908" Mar 17 09:17:39 crc kubenswrapper[4813]: I0317 09:17:39.856161 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-54j2x" podStartSLOduration=3.14222577 podStartE2EDuration="5.856151732s" podCreationTimestamp="2026-03-17 09:17:34 +0000 UTC" firstStartedPulling="2026-03-17 09:17:35.7605307 +0000 UTC m=+477.861334199" lastFinishedPulling="2026-03-17 09:17:38.474456652 +0000 UTC m=+480.575260161" observedRunningTime="2026-03-17 09:17:38.841543969 +0000 UTC m=+480.942347478" watchObservedRunningTime="2026-03-17 09:17:39.856151732 +0000 UTC m=+481.956955231" Mar 17 09:17:40 crc kubenswrapper[4813]: I0317 09:17:40.838908 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cm7vw" event={"ID":"ba5c8bcc-1783-4889-a504-4233deca14dd","Type":"ContainerStarted","Data":"3b53c27ab24c90a649a1c219348d70d9599a4fdb36136d288a39b0ac47e28c8b"} Mar 17 09:17:40 crc kubenswrapper[4813]: I0317 09:17:40.863723 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-cm7vw" podStartSLOduration=2.301088329 podStartE2EDuration="4.863708147s" podCreationTimestamp="2026-03-17 09:17:36 +0000 UTC" firstStartedPulling="2026-03-17 09:17:37.806146274 +0000 UTC m=+479.906949783" lastFinishedPulling="2026-03-17 09:17:40.368766102 +0000 UTC m=+482.469569601" observedRunningTime="2026-03-17 09:17:40.860138588 +0000 UTC m=+482.960942087" watchObservedRunningTime="2026-03-17 09:17:40.863708147 +0000 UTC m=+482.964511636" Mar 17 09:17:43 crc kubenswrapper[4813]: I0317 09:17:43.083297 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rshd8" Mar 17 09:17:43 crc kubenswrapper[4813]: I0317 09:17:43.083783 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rshd8" Mar 17 09:17:43 crc kubenswrapper[4813]: I0317 09:17:43.142811 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rshd8" Mar 17 09:17:43 crc kubenswrapper[4813]: I0317 09:17:43.919818 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rshd8" Mar 17 09:17:44 crc kubenswrapper[4813]: I0317 09:17:44.114459 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:17:44 crc kubenswrapper[4813]: I0317 09:17:44.114520 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:17:44 crc kubenswrapper[4813]: I0317 09:17:44.212941 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-j7m85" Mar 17 09:17:44 crc kubenswrapper[4813]: I0317 09:17:44.213000 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-j7m85" Mar 17 09:17:44 crc kubenswrapper[4813]: I0317 09:17:44.259353 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-j7m85" Mar 17 09:17:44 crc kubenswrapper[4813]: I0317 09:17:44.923254 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-j7m85" Mar 17 09:17:45 crc kubenswrapper[4813]: I0317 09:17:45.241140 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-54j2x" Mar 17 09:17:45 crc kubenswrapper[4813]: I0317 09:17:45.242511 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-54j2x" Mar 17 09:17:45 crc kubenswrapper[4813]: I0317 09:17:45.311544 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-54j2x" Mar 17 09:17:45 crc kubenswrapper[4813]: I0317 09:17:45.957655 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-54j2x" Mar 17 09:17:46 crc kubenswrapper[4813]: I0317 09:17:46.618007 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-cm7vw" Mar 17 09:17:46 crc kubenswrapper[4813]: I0317 09:17:46.618102 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-cm7vw" Mar 17 09:17:47 crc kubenswrapper[4813]: I0317 09:17:47.686404 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-cm7vw" podUID="ba5c8bcc-1783-4889-a504-4233deca14dd" containerName="registry-server" probeResult="failure" output=< Mar 17 09:17:47 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Mar 17 09:17:47 crc kubenswrapper[4813]: > Mar 17 09:17:56 crc kubenswrapper[4813]: I0317 09:17:56.668337 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-cm7vw" Mar 17 09:17:56 crc kubenswrapper[4813]: I0317 09:17:56.727965 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-cm7vw" Mar 17 09:18:00 crc kubenswrapper[4813]: I0317 09:18:00.143267 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562318-l68gh"] Mar 17 09:18:00 crc kubenswrapper[4813]: I0317 09:18:00.144307 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562318-l68gh" Mar 17 09:18:00 crc kubenswrapper[4813]: I0317 09:18:00.146742 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:18:00 crc kubenswrapper[4813]: I0317 09:18:00.147104 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:18:00 crc kubenswrapper[4813]: I0317 09:18:00.147522 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:18:00 crc kubenswrapper[4813]: I0317 09:18:00.152822 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562318-l68gh"] Mar 17 09:18:00 crc kubenswrapper[4813]: I0317 09:18:00.267370 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z54n\" (UniqueName: \"kubernetes.io/projected/5b2354e2-d41f-47f9-9fd9-11f51b893a3a-kube-api-access-8z54n\") pod \"auto-csr-approver-29562318-l68gh\" (UID: \"5b2354e2-d41f-47f9-9fd9-11f51b893a3a\") " pod="openshift-infra/auto-csr-approver-29562318-l68gh" Mar 17 09:18:00 crc kubenswrapper[4813]: I0317 09:18:00.369248 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z54n\" (UniqueName: \"kubernetes.io/projected/5b2354e2-d41f-47f9-9fd9-11f51b893a3a-kube-api-access-8z54n\") pod \"auto-csr-approver-29562318-l68gh\" (UID: \"5b2354e2-d41f-47f9-9fd9-11f51b893a3a\") " pod="openshift-infra/auto-csr-approver-29562318-l68gh" Mar 17 09:18:00 crc kubenswrapper[4813]: I0317 09:18:00.389842 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z54n\" (UniqueName: \"kubernetes.io/projected/5b2354e2-d41f-47f9-9fd9-11f51b893a3a-kube-api-access-8z54n\") pod \"auto-csr-approver-29562318-l68gh\" (UID: \"5b2354e2-d41f-47f9-9fd9-11f51b893a3a\") " pod="openshift-infra/auto-csr-approver-29562318-l68gh" Mar 17 09:18:00 crc kubenswrapper[4813]: I0317 09:18:00.464015 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562318-l68gh" Mar 17 09:18:00 crc kubenswrapper[4813]: I0317 09:18:00.890679 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562318-l68gh"] Mar 17 09:18:00 crc kubenswrapper[4813]: I0317 09:18:00.961947 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562318-l68gh" event={"ID":"5b2354e2-d41f-47f9-9fd9-11f51b893a3a","Type":"ContainerStarted","Data":"685018a9dda4e6e71dc6b0bde5e504ed00fb22a5bd1246343173e82a3d605867"} Mar 17 09:18:01 crc kubenswrapper[4813]: I0317 09:18:01.734179 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" podUID="782b4936-2caa-44da-8716-c57924be6df3" containerName="registry" containerID="cri-o://be5066153cadae7bb93f3adaacd29a207801bea4da3722fc5f08063db6f705c8" gracePeriod=30 Mar 17 09:18:01 crc kubenswrapper[4813]: I0317 09:18:01.968616 4813 generic.go:334] "Generic (PLEG): container finished" podID="782b4936-2caa-44da-8716-c57924be6df3" containerID="be5066153cadae7bb93f3adaacd29a207801bea4da3722fc5f08063db6f705c8" exitCode=0 Mar 17 09:18:01 crc kubenswrapper[4813]: I0317 09:18:01.968662 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" event={"ID":"782b4936-2caa-44da-8716-c57924be6df3","Type":"ContainerDied","Data":"be5066153cadae7bb93f3adaacd29a207801bea4da3722fc5f08063db6f705c8"} Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.163385 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.300203 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/782b4936-2caa-44da-8716-c57924be6df3-registry-certificates\") pod \"782b4936-2caa-44da-8716-c57924be6df3\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.300253 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/782b4936-2caa-44da-8716-c57924be6df3-trusted-ca\") pod \"782b4936-2caa-44da-8716-c57924be6df3\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.300382 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"782b4936-2caa-44da-8716-c57924be6df3\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.300425 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/782b4936-2caa-44da-8716-c57924be6df3-bound-sa-token\") pod \"782b4936-2caa-44da-8716-c57924be6df3\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.300458 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/782b4936-2caa-44da-8716-c57924be6df3-installation-pull-secrets\") pod \"782b4936-2caa-44da-8716-c57924be6df3\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.300493 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/782b4936-2caa-44da-8716-c57924be6df3-registry-tls\") pod \"782b4936-2caa-44da-8716-c57924be6df3\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.300510 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kdnwc\" (UniqueName: \"kubernetes.io/projected/782b4936-2caa-44da-8716-c57924be6df3-kube-api-access-kdnwc\") pod \"782b4936-2caa-44da-8716-c57924be6df3\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.300528 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/782b4936-2caa-44da-8716-c57924be6df3-ca-trust-extracted\") pod \"782b4936-2caa-44da-8716-c57924be6df3\" (UID: \"782b4936-2caa-44da-8716-c57924be6df3\") " Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.301503 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/782b4936-2caa-44da-8716-c57924be6df3-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "782b4936-2caa-44da-8716-c57924be6df3" (UID: "782b4936-2caa-44da-8716-c57924be6df3"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.301767 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/782b4936-2caa-44da-8716-c57924be6df3-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "782b4936-2caa-44da-8716-c57924be6df3" (UID: "782b4936-2caa-44da-8716-c57924be6df3"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.311373 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "782b4936-2caa-44da-8716-c57924be6df3" (UID: "782b4936-2caa-44da-8716-c57924be6df3"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.313055 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/782b4936-2caa-44da-8716-c57924be6df3-kube-api-access-kdnwc" (OuterVolumeSpecName: "kube-api-access-kdnwc") pod "782b4936-2caa-44da-8716-c57924be6df3" (UID: "782b4936-2caa-44da-8716-c57924be6df3"). InnerVolumeSpecName "kube-api-access-kdnwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.313057 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/782b4936-2caa-44da-8716-c57924be6df3-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "782b4936-2caa-44da-8716-c57924be6df3" (UID: "782b4936-2caa-44da-8716-c57924be6df3"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.313251 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/782b4936-2caa-44da-8716-c57924be6df3-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "782b4936-2caa-44da-8716-c57924be6df3" (UID: "782b4936-2caa-44da-8716-c57924be6df3"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.316245 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/782b4936-2caa-44da-8716-c57924be6df3-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "782b4936-2caa-44da-8716-c57924be6df3" (UID: "782b4936-2caa-44da-8716-c57924be6df3"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.317432 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/782b4936-2caa-44da-8716-c57924be6df3-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "782b4936-2caa-44da-8716-c57924be6df3" (UID: "782b4936-2caa-44da-8716-c57924be6df3"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.401950 4813 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/782b4936-2caa-44da-8716-c57924be6df3-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.401984 4813 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/782b4936-2caa-44da-8716-c57924be6df3-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.401999 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kdnwc\" (UniqueName: \"kubernetes.io/projected/782b4936-2caa-44da-8716-c57924be6df3-kube-api-access-kdnwc\") on node \"crc\" DevicePath \"\"" Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.402012 4813 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/782b4936-2caa-44da-8716-c57924be6df3-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.402141 4813 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/782b4936-2caa-44da-8716-c57924be6df3-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.402152 4813 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/782b4936-2caa-44da-8716-c57924be6df3-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.402162 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/782b4936-2caa-44da-8716-c57924be6df3-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.975516 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" event={"ID":"782b4936-2caa-44da-8716-c57924be6df3","Type":"ContainerDied","Data":"0c096ac2c60f7bd51d4760d2758d8e014342c7765f19a533f26c9d79d8918b75"} Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.975899 4813 scope.go:117] "RemoveContainer" containerID="be5066153cadae7bb93f3adaacd29a207801bea4da3722fc5f08063db6f705c8" Mar 17 09:18:02 crc kubenswrapper[4813]: I0317 09:18:02.976090 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-dbm4b" Mar 17 09:18:03 crc kubenswrapper[4813]: I0317 09:18:03.572018 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dbm4b"] Mar 17 09:18:03 crc kubenswrapper[4813]: I0317 09:18:03.576353 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dbm4b"] Mar 17 09:18:03 crc kubenswrapper[4813]: I0317 09:18:03.983083 4813 generic.go:334] "Generic (PLEG): container finished" podID="5b2354e2-d41f-47f9-9fd9-11f51b893a3a" containerID="dad746fa90ef589dad6ffc30fda9d259a6ebb043370f2f9ce8eff8745df59eb2" exitCode=0 Mar 17 09:18:03 crc kubenswrapper[4813]: I0317 09:18:03.983196 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562318-l68gh" event={"ID":"5b2354e2-d41f-47f9-9fd9-11f51b893a3a","Type":"ContainerDied","Data":"dad746fa90ef589dad6ffc30fda9d259a6ebb043370f2f9ce8eff8745df59eb2"} Mar 17 09:18:04 crc kubenswrapper[4813]: I0317 09:18:04.746009 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="782b4936-2caa-44da-8716-c57924be6df3" path="/var/lib/kubelet/pods/782b4936-2caa-44da-8716-c57924be6df3/volumes" Mar 17 09:18:05 crc kubenswrapper[4813]: I0317 09:18:05.240848 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562318-l68gh" Mar 17 09:18:05 crc kubenswrapper[4813]: I0317 09:18:05.354792 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8z54n\" (UniqueName: \"kubernetes.io/projected/5b2354e2-d41f-47f9-9fd9-11f51b893a3a-kube-api-access-8z54n\") pod \"5b2354e2-d41f-47f9-9fd9-11f51b893a3a\" (UID: \"5b2354e2-d41f-47f9-9fd9-11f51b893a3a\") " Mar 17 09:18:05 crc kubenswrapper[4813]: I0317 09:18:05.360801 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b2354e2-d41f-47f9-9fd9-11f51b893a3a-kube-api-access-8z54n" (OuterVolumeSpecName: "kube-api-access-8z54n") pod "5b2354e2-d41f-47f9-9fd9-11f51b893a3a" (UID: "5b2354e2-d41f-47f9-9fd9-11f51b893a3a"). InnerVolumeSpecName "kube-api-access-8z54n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:18:05 crc kubenswrapper[4813]: I0317 09:18:05.456259 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8z54n\" (UniqueName: \"kubernetes.io/projected/5b2354e2-d41f-47f9-9fd9-11f51b893a3a-kube-api-access-8z54n\") on node \"crc\" DevicePath \"\"" Mar 17 09:18:05 crc kubenswrapper[4813]: I0317 09:18:05.996677 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562318-l68gh" event={"ID":"5b2354e2-d41f-47f9-9fd9-11f51b893a3a","Type":"ContainerDied","Data":"685018a9dda4e6e71dc6b0bde5e504ed00fb22a5bd1246343173e82a3d605867"} Mar 17 09:18:05 crc kubenswrapper[4813]: I0317 09:18:05.997011 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="685018a9dda4e6e71dc6b0bde5e504ed00fb22a5bd1246343173e82a3d605867" Mar 17 09:18:05 crc kubenswrapper[4813]: I0317 09:18:05.996724 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562318-l68gh" Mar 17 09:18:06 crc kubenswrapper[4813]: I0317 09:18:06.299977 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562312-hjpcp"] Mar 17 09:18:06 crc kubenswrapper[4813]: I0317 09:18:06.305813 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562312-hjpcp"] Mar 17 09:18:06 crc kubenswrapper[4813]: I0317 09:18:06.736577 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8626af43-6b5e-49e1-b19f-3986cdecdd2d" path="/var/lib/kubelet/pods/8626af43-6b5e-49e1-b19f-3986cdecdd2d/volumes" Mar 17 09:18:14 crc kubenswrapper[4813]: I0317 09:18:14.113925 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:18:14 crc kubenswrapper[4813]: I0317 09:18:14.114704 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:18:14 crc kubenswrapper[4813]: I0317 09:18:14.114746 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:18:14 crc kubenswrapper[4813]: I0317 09:18:14.115344 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"af70be0dbae35ba7879eb8d4732f829e536191eac3391f7a3a43fe520584a7bf"} pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 09:18:14 crc kubenswrapper[4813]: I0317 09:18:14.115400 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" containerID="cri-o://af70be0dbae35ba7879eb8d4732f829e536191eac3391f7a3a43fe520584a7bf" gracePeriod=600 Mar 17 09:18:15 crc kubenswrapper[4813]: I0317 09:18:15.047888 4813 generic.go:334] "Generic (PLEG): container finished" podID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerID="af70be0dbae35ba7879eb8d4732f829e536191eac3391f7a3a43fe520584a7bf" exitCode=0 Mar 17 09:18:15 crc kubenswrapper[4813]: I0317 09:18:15.047996 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerDied","Data":"af70be0dbae35ba7879eb8d4732f829e536191eac3391f7a3a43fe520584a7bf"} Mar 17 09:18:15 crc kubenswrapper[4813]: I0317 09:18:15.048153 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerStarted","Data":"e0ba14bb710e8ef535a87d43e489a6972a73ab00b1733aa05cbdc8da2506f48b"} Mar 17 09:18:15 crc kubenswrapper[4813]: I0317 09:18:15.048174 4813 scope.go:117] "RemoveContainer" containerID="1e56c8da97c337dc7e423300154c5d557de77d5d35a638ded9ca14000551bd09" Mar 17 09:18:33 crc kubenswrapper[4813]: I0317 09:18:33.547371 4813 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","burstable","pod782b4936-2caa-44da-8716-c57924be6df3"] err="unable to destroy cgroup paths for cgroup [kubepods burstable pod782b4936-2caa-44da-8716-c57924be6df3] : Timed out while waiting for systemd to remove kubepods-burstable-pod782b4936_2caa_44da_8716_c57924be6df3.slice" Mar 17 09:20:00 crc kubenswrapper[4813]: I0317 09:20:00.143474 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562320-bfknz"] Mar 17 09:20:00 crc kubenswrapper[4813]: E0317 09:20:00.144394 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="782b4936-2caa-44da-8716-c57924be6df3" containerName="registry" Mar 17 09:20:00 crc kubenswrapper[4813]: I0317 09:20:00.144415 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="782b4936-2caa-44da-8716-c57924be6df3" containerName="registry" Mar 17 09:20:00 crc kubenswrapper[4813]: E0317 09:20:00.144444 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b2354e2-d41f-47f9-9fd9-11f51b893a3a" containerName="oc" Mar 17 09:20:00 crc kubenswrapper[4813]: I0317 09:20:00.144456 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b2354e2-d41f-47f9-9fd9-11f51b893a3a" containerName="oc" Mar 17 09:20:00 crc kubenswrapper[4813]: I0317 09:20:00.144667 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b2354e2-d41f-47f9-9fd9-11f51b893a3a" containerName="oc" Mar 17 09:20:00 crc kubenswrapper[4813]: I0317 09:20:00.144687 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="782b4936-2caa-44da-8716-c57924be6df3" containerName="registry" Mar 17 09:20:00 crc kubenswrapper[4813]: I0317 09:20:00.145252 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562320-bfknz" Mar 17 09:20:00 crc kubenswrapper[4813]: I0317 09:20:00.150271 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:20:00 crc kubenswrapper[4813]: I0317 09:20:00.150443 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:20:00 crc kubenswrapper[4813]: I0317 09:20:00.150799 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:20:00 crc kubenswrapper[4813]: I0317 09:20:00.156265 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562320-bfknz"] Mar 17 09:20:00 crc kubenswrapper[4813]: I0317 09:20:00.239137 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq7k2\" (UniqueName: \"kubernetes.io/projected/75bf03ea-aeae-4e4b-b774-e4f1b45981d2-kube-api-access-xq7k2\") pod \"auto-csr-approver-29562320-bfknz\" (UID: \"75bf03ea-aeae-4e4b-b774-e4f1b45981d2\") " pod="openshift-infra/auto-csr-approver-29562320-bfknz" Mar 17 09:20:00 crc kubenswrapper[4813]: I0317 09:20:00.341188 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq7k2\" (UniqueName: \"kubernetes.io/projected/75bf03ea-aeae-4e4b-b774-e4f1b45981d2-kube-api-access-xq7k2\") pod \"auto-csr-approver-29562320-bfknz\" (UID: \"75bf03ea-aeae-4e4b-b774-e4f1b45981d2\") " pod="openshift-infra/auto-csr-approver-29562320-bfknz" Mar 17 09:20:00 crc kubenswrapper[4813]: I0317 09:20:00.370737 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq7k2\" (UniqueName: \"kubernetes.io/projected/75bf03ea-aeae-4e4b-b774-e4f1b45981d2-kube-api-access-xq7k2\") pod \"auto-csr-approver-29562320-bfknz\" (UID: \"75bf03ea-aeae-4e4b-b774-e4f1b45981d2\") " pod="openshift-infra/auto-csr-approver-29562320-bfknz" Mar 17 09:20:00 crc kubenswrapper[4813]: I0317 09:20:00.479434 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562320-bfknz" Mar 17 09:20:00 crc kubenswrapper[4813]: I0317 09:20:00.751326 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562320-bfknz"] Mar 17 09:20:00 crc kubenswrapper[4813]: I0317 09:20:00.765257 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 09:20:00 crc kubenswrapper[4813]: I0317 09:20:00.798908 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562320-bfknz" event={"ID":"75bf03ea-aeae-4e4b-b774-e4f1b45981d2","Type":"ContainerStarted","Data":"339abeaf8eb8df1350f62dc4ae0221eb74e90a05352acf328fa21f623944e1fe"} Mar 17 09:20:02 crc kubenswrapper[4813]: I0317 09:20:02.818789 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562320-bfknz" event={"ID":"75bf03ea-aeae-4e4b-b774-e4f1b45981d2","Type":"ContainerStarted","Data":"14e897629c918331483db943441831f8093d0c3c781e5e96d5e251f70a2a8e54"} Mar 17 09:20:02 crc kubenswrapper[4813]: I0317 09:20:02.838279 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562320-bfknz" podStartSLOduration=1.228553243 podStartE2EDuration="2.838260396s" podCreationTimestamp="2026-03-17 09:20:00 +0000 UTC" firstStartedPulling="2026-03-17 09:20:00.765021815 +0000 UTC m=+622.865825314" lastFinishedPulling="2026-03-17 09:20:02.374728958 +0000 UTC m=+624.475532467" observedRunningTime="2026-03-17 09:20:02.833855675 +0000 UTC m=+624.934659184" watchObservedRunningTime="2026-03-17 09:20:02.838260396 +0000 UTC m=+624.939063895" Mar 17 09:20:03 crc kubenswrapper[4813]: I0317 09:20:03.826885 4813 generic.go:334] "Generic (PLEG): container finished" podID="75bf03ea-aeae-4e4b-b774-e4f1b45981d2" containerID="14e897629c918331483db943441831f8093d0c3c781e5e96d5e251f70a2a8e54" exitCode=0 Mar 17 09:20:03 crc kubenswrapper[4813]: I0317 09:20:03.826945 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562320-bfknz" event={"ID":"75bf03ea-aeae-4e4b-b774-e4f1b45981d2","Type":"ContainerDied","Data":"14e897629c918331483db943441831f8093d0c3c781e5e96d5e251f70a2a8e54"} Mar 17 09:20:05 crc kubenswrapper[4813]: I0317 09:20:05.145846 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562320-bfknz" Mar 17 09:20:05 crc kubenswrapper[4813]: I0317 09:20:05.234704 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xq7k2\" (UniqueName: \"kubernetes.io/projected/75bf03ea-aeae-4e4b-b774-e4f1b45981d2-kube-api-access-xq7k2\") pod \"75bf03ea-aeae-4e4b-b774-e4f1b45981d2\" (UID: \"75bf03ea-aeae-4e4b-b774-e4f1b45981d2\") " Mar 17 09:20:05 crc kubenswrapper[4813]: I0317 09:20:05.247919 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75bf03ea-aeae-4e4b-b774-e4f1b45981d2-kube-api-access-xq7k2" (OuterVolumeSpecName: "kube-api-access-xq7k2") pod "75bf03ea-aeae-4e4b-b774-e4f1b45981d2" (UID: "75bf03ea-aeae-4e4b-b774-e4f1b45981d2"). InnerVolumeSpecName "kube-api-access-xq7k2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:20:05 crc kubenswrapper[4813]: I0317 09:20:05.336300 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xq7k2\" (UniqueName: \"kubernetes.io/projected/75bf03ea-aeae-4e4b-b774-e4f1b45981d2-kube-api-access-xq7k2\") on node \"crc\" DevicePath \"\"" Mar 17 09:20:05 crc kubenswrapper[4813]: I0317 09:20:05.850954 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562320-bfknz" event={"ID":"75bf03ea-aeae-4e4b-b774-e4f1b45981d2","Type":"ContainerDied","Data":"339abeaf8eb8df1350f62dc4ae0221eb74e90a05352acf328fa21f623944e1fe"} Mar 17 09:20:05 crc kubenswrapper[4813]: I0317 09:20:05.851011 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="339abeaf8eb8df1350f62dc4ae0221eb74e90a05352acf328fa21f623944e1fe" Mar 17 09:20:05 crc kubenswrapper[4813]: I0317 09:20:05.851106 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562320-bfknz" Mar 17 09:20:05 crc kubenswrapper[4813]: I0317 09:20:05.915549 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562314-dwd8b"] Mar 17 09:20:05 crc kubenswrapper[4813]: I0317 09:20:05.922729 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562314-dwd8b"] Mar 17 09:20:06 crc kubenswrapper[4813]: I0317 09:20:06.742088 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9225f4aa-611e-447c-94b7-804e8c973203" path="/var/lib/kubelet/pods/9225f4aa-611e-447c-94b7-804e8c973203/volumes" Mar 17 09:20:14 crc kubenswrapper[4813]: I0317 09:20:14.114256 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:20:14 crc kubenswrapper[4813]: I0317 09:20:14.114921 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:20:39 crc kubenswrapper[4813]: I0317 09:20:39.086473 4813 scope.go:117] "RemoveContainer" containerID="eb525fa6ddd5e7d38d337ed03384d69605ce16132833739caba5ba0ad5e4edab" Mar 17 09:20:39 crc kubenswrapper[4813]: I0317 09:20:39.140322 4813 scope.go:117] "RemoveContainer" containerID="489850af96666df4ef1e82730a50eb36058eea72a391bebec30181a812172f63" Mar 17 09:20:44 crc kubenswrapper[4813]: I0317 09:20:44.114099 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:20:44 crc kubenswrapper[4813]: I0317 09:20:44.114176 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:21:14 crc kubenswrapper[4813]: I0317 09:21:14.114513 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:21:14 crc kubenswrapper[4813]: I0317 09:21:14.115236 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:21:14 crc kubenswrapper[4813]: I0317 09:21:14.115314 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:21:14 crc kubenswrapper[4813]: I0317 09:21:14.116241 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e0ba14bb710e8ef535a87d43e489a6972a73ab00b1733aa05cbdc8da2506f48b"} pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 09:21:14 crc kubenswrapper[4813]: I0317 09:21:14.116330 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" containerID="cri-o://e0ba14bb710e8ef535a87d43e489a6972a73ab00b1733aa05cbdc8da2506f48b" gracePeriod=600 Mar 17 09:21:14 crc kubenswrapper[4813]: I0317 09:21:14.339769 4813 generic.go:334] "Generic (PLEG): container finished" podID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerID="e0ba14bb710e8ef535a87d43e489a6972a73ab00b1733aa05cbdc8da2506f48b" exitCode=0 Mar 17 09:21:14 crc kubenswrapper[4813]: I0317 09:21:14.339851 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerDied","Data":"e0ba14bb710e8ef535a87d43e489a6972a73ab00b1733aa05cbdc8da2506f48b"} Mar 17 09:21:14 crc kubenswrapper[4813]: I0317 09:21:14.340075 4813 scope.go:117] "RemoveContainer" containerID="af70be0dbae35ba7879eb8d4732f829e536191eac3391f7a3a43fe520584a7bf" Mar 17 09:21:15 crc kubenswrapper[4813]: I0317 09:21:15.353370 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerStarted","Data":"b67bb2399345967e5d98c3e93bfb13d1f15014e04cd83984dc098f8a87b2a52c"} Mar 17 09:22:00 crc kubenswrapper[4813]: I0317 09:22:00.145285 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562322-q8wzc"] Mar 17 09:22:00 crc kubenswrapper[4813]: E0317 09:22:00.146277 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75bf03ea-aeae-4e4b-b774-e4f1b45981d2" containerName="oc" Mar 17 09:22:00 crc kubenswrapper[4813]: I0317 09:22:00.146306 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="75bf03ea-aeae-4e4b-b774-e4f1b45981d2" containerName="oc" Mar 17 09:22:00 crc kubenswrapper[4813]: I0317 09:22:00.146516 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="75bf03ea-aeae-4e4b-b774-e4f1b45981d2" containerName="oc" Mar 17 09:22:00 crc kubenswrapper[4813]: I0317 09:22:00.147137 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562322-q8wzc" Mar 17 09:22:00 crc kubenswrapper[4813]: I0317 09:22:00.149577 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:22:00 crc kubenswrapper[4813]: I0317 09:22:00.149891 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:22:00 crc kubenswrapper[4813]: I0317 09:22:00.149981 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:22:00 crc kubenswrapper[4813]: I0317 09:22:00.156768 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562322-q8wzc"] Mar 17 09:22:00 crc kubenswrapper[4813]: I0317 09:22:00.345993 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46bms\" (UniqueName: \"kubernetes.io/projected/6a3fed28-e266-47a5-851f-f04fb94804e8-kube-api-access-46bms\") pod \"auto-csr-approver-29562322-q8wzc\" (UID: \"6a3fed28-e266-47a5-851f-f04fb94804e8\") " pod="openshift-infra/auto-csr-approver-29562322-q8wzc" Mar 17 09:22:00 crc kubenswrapper[4813]: I0317 09:22:00.457159 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46bms\" (UniqueName: \"kubernetes.io/projected/6a3fed28-e266-47a5-851f-f04fb94804e8-kube-api-access-46bms\") pod \"auto-csr-approver-29562322-q8wzc\" (UID: \"6a3fed28-e266-47a5-851f-f04fb94804e8\") " pod="openshift-infra/auto-csr-approver-29562322-q8wzc" Mar 17 09:22:00 crc kubenswrapper[4813]: I0317 09:22:00.493793 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46bms\" (UniqueName: \"kubernetes.io/projected/6a3fed28-e266-47a5-851f-f04fb94804e8-kube-api-access-46bms\") pod \"auto-csr-approver-29562322-q8wzc\" (UID: \"6a3fed28-e266-47a5-851f-f04fb94804e8\") " pod="openshift-infra/auto-csr-approver-29562322-q8wzc" Mar 17 09:22:00 crc kubenswrapper[4813]: I0317 09:22:00.775717 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562322-q8wzc" Mar 17 09:22:00 crc kubenswrapper[4813]: I0317 09:22:00.975490 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562322-q8wzc"] Mar 17 09:22:00 crc kubenswrapper[4813]: W0317 09:22:00.978559 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a3fed28_e266_47a5_851f_f04fb94804e8.slice/crio-19f281e4cd4509dadd5d188dcd754a010f0e5500484fef2257f911aa3a6ea7b3 WatchSource:0}: Error finding container 19f281e4cd4509dadd5d188dcd754a010f0e5500484fef2257f911aa3a6ea7b3: Status 404 returned error can't find the container with id 19f281e4cd4509dadd5d188dcd754a010f0e5500484fef2257f911aa3a6ea7b3 Mar 17 09:22:01 crc kubenswrapper[4813]: I0317 09:22:01.686762 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562322-q8wzc" event={"ID":"6a3fed28-e266-47a5-851f-f04fb94804e8","Type":"ContainerStarted","Data":"19f281e4cd4509dadd5d188dcd754a010f0e5500484fef2257f911aa3a6ea7b3"} Mar 17 09:22:02 crc kubenswrapper[4813]: I0317 09:22:02.696542 4813 generic.go:334] "Generic (PLEG): container finished" podID="6a3fed28-e266-47a5-851f-f04fb94804e8" containerID="7213f79e91ecdab9ed18fc3a07525828594cef72a9583cc67db80f37da5f05eb" exitCode=0 Mar 17 09:22:02 crc kubenswrapper[4813]: I0317 09:22:02.696610 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562322-q8wzc" event={"ID":"6a3fed28-e266-47a5-851f-f04fb94804e8","Type":"ContainerDied","Data":"7213f79e91ecdab9ed18fc3a07525828594cef72a9583cc67db80f37da5f05eb"} Mar 17 09:22:04 crc kubenswrapper[4813]: I0317 09:22:04.036025 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562322-q8wzc" Mar 17 09:22:04 crc kubenswrapper[4813]: I0317 09:22:04.105245 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-46bms\" (UniqueName: \"kubernetes.io/projected/6a3fed28-e266-47a5-851f-f04fb94804e8-kube-api-access-46bms\") pod \"6a3fed28-e266-47a5-851f-f04fb94804e8\" (UID: \"6a3fed28-e266-47a5-851f-f04fb94804e8\") " Mar 17 09:22:04 crc kubenswrapper[4813]: I0317 09:22:04.110733 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a3fed28-e266-47a5-851f-f04fb94804e8-kube-api-access-46bms" (OuterVolumeSpecName: "kube-api-access-46bms") pod "6a3fed28-e266-47a5-851f-f04fb94804e8" (UID: "6a3fed28-e266-47a5-851f-f04fb94804e8"). InnerVolumeSpecName "kube-api-access-46bms". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:22:04 crc kubenswrapper[4813]: I0317 09:22:04.206949 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-46bms\" (UniqueName: \"kubernetes.io/projected/6a3fed28-e266-47a5-851f-f04fb94804e8-kube-api-access-46bms\") on node \"crc\" DevicePath \"\"" Mar 17 09:22:04 crc kubenswrapper[4813]: I0317 09:22:04.712400 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562322-q8wzc" event={"ID":"6a3fed28-e266-47a5-851f-f04fb94804e8","Type":"ContainerDied","Data":"19f281e4cd4509dadd5d188dcd754a010f0e5500484fef2257f911aa3a6ea7b3"} Mar 17 09:22:04 crc kubenswrapper[4813]: I0317 09:22:04.712439 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562322-q8wzc" Mar 17 09:22:04 crc kubenswrapper[4813]: I0317 09:22:04.712457 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19f281e4cd4509dadd5d188dcd754a010f0e5500484fef2257f911aa3a6ea7b3" Mar 17 09:22:05 crc kubenswrapper[4813]: I0317 09:22:05.112354 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562316-rzldf"] Mar 17 09:22:05 crc kubenswrapper[4813]: I0317 09:22:05.118725 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562316-rzldf"] Mar 17 09:22:06 crc kubenswrapper[4813]: I0317 09:22:06.741738 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="86b710cc-c717-4373-8b26-89d1d6d66bd6" path="/var/lib/kubelet/pods/86b710cc-c717-4373-8b26-89d1d6d66bd6/volumes" Mar 17 09:22:39 crc kubenswrapper[4813]: I0317 09:22:39.236157 4813 scope.go:117] "RemoveContainer" containerID="7ccdaaae72fb5d598178037112459ee1588a87f04acdb9b6eb7c0cfcfa485397" Mar 17 09:23:14 crc kubenswrapper[4813]: I0317 09:23:14.114051 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:23:14 crc kubenswrapper[4813]: I0317 09:23:14.114923 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:23:44 crc kubenswrapper[4813]: I0317 09:23:44.114376 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:23:44 crc kubenswrapper[4813]: I0317 09:23:44.115187 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.316889 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-slrw5"] Mar 17 09:23:45 crc kubenswrapper[4813]: E0317 09:23:45.317090 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a3fed28-e266-47a5-851f-f04fb94804e8" containerName="oc" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.317101 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a3fed28-e266-47a5-851f-f04fb94804e8" containerName="oc" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.317185 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a3fed28-e266-47a5-851f-f04fb94804e8" containerName="oc" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.317512 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-slrw5" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.319818 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.320581 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.320582 4813 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-rlgbt" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.324252 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dfwwv\" (UniqueName: \"kubernetes.io/projected/7ff1e0fd-1a34-4f70-97c0-9fbb47f0b52b-kube-api-access-dfwwv\") pod \"cert-manager-cainjector-cf98fcc89-slrw5\" (UID: \"7ff1e0fd-1a34-4f70-97c0-9fbb47f0b52b\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-slrw5" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.343941 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-858654f9db-knkd2"] Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.344571 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-knkd2" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.348004 4813 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-cqp5s" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.350107 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-slrw5"] Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.356427 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-hdhpn"] Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.357111 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-hdhpn" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.358886 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-knkd2"] Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.359081 4813 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-thfrw" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.372408 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-hdhpn"] Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.424900 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxtr8\" (UniqueName: \"kubernetes.io/projected/d949287d-f3e6-4ac8-8c16-a0d0652ae302-kube-api-access-pxtr8\") pod \"cert-manager-webhook-687f57d79b-hdhpn\" (UID: \"d949287d-f3e6-4ac8-8c16-a0d0652ae302\") " pod="cert-manager/cert-manager-webhook-687f57d79b-hdhpn" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.424947 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgs47\" (UniqueName: \"kubernetes.io/projected/08c8f504-aa0d-47b9-ab86-a4b440c6f446-kube-api-access-qgs47\") pod \"cert-manager-858654f9db-knkd2\" (UID: \"08c8f504-aa0d-47b9-ab86-a4b440c6f446\") " pod="cert-manager/cert-manager-858654f9db-knkd2" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.424970 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dfwwv\" (UniqueName: \"kubernetes.io/projected/7ff1e0fd-1a34-4f70-97c0-9fbb47f0b52b-kube-api-access-dfwwv\") pod \"cert-manager-cainjector-cf98fcc89-slrw5\" (UID: \"7ff1e0fd-1a34-4f70-97c0-9fbb47f0b52b\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-slrw5" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.441138 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dfwwv\" (UniqueName: \"kubernetes.io/projected/7ff1e0fd-1a34-4f70-97c0-9fbb47f0b52b-kube-api-access-dfwwv\") pod \"cert-manager-cainjector-cf98fcc89-slrw5\" (UID: \"7ff1e0fd-1a34-4f70-97c0-9fbb47f0b52b\") " pod="cert-manager/cert-manager-cainjector-cf98fcc89-slrw5" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.525329 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxtr8\" (UniqueName: \"kubernetes.io/projected/d949287d-f3e6-4ac8-8c16-a0d0652ae302-kube-api-access-pxtr8\") pod \"cert-manager-webhook-687f57d79b-hdhpn\" (UID: \"d949287d-f3e6-4ac8-8c16-a0d0652ae302\") " pod="cert-manager/cert-manager-webhook-687f57d79b-hdhpn" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.525628 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgs47\" (UniqueName: \"kubernetes.io/projected/08c8f504-aa0d-47b9-ab86-a4b440c6f446-kube-api-access-qgs47\") pod \"cert-manager-858654f9db-knkd2\" (UID: \"08c8f504-aa0d-47b9-ab86-a4b440c6f446\") " pod="cert-manager/cert-manager-858654f9db-knkd2" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.544064 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxtr8\" (UniqueName: \"kubernetes.io/projected/d949287d-f3e6-4ac8-8c16-a0d0652ae302-kube-api-access-pxtr8\") pod \"cert-manager-webhook-687f57d79b-hdhpn\" (UID: \"d949287d-f3e6-4ac8-8c16-a0d0652ae302\") " pod="cert-manager/cert-manager-webhook-687f57d79b-hdhpn" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.544318 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgs47\" (UniqueName: \"kubernetes.io/projected/08c8f504-aa0d-47b9-ab86-a4b440c6f446-kube-api-access-qgs47\") pod \"cert-manager-858654f9db-knkd2\" (UID: \"08c8f504-aa0d-47b9-ab86-a4b440c6f446\") " pod="cert-manager/cert-manager-858654f9db-knkd2" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.641938 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-cf98fcc89-slrw5" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.660840 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-858654f9db-knkd2" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.671546 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-687f57d79b-hdhpn" Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.923504 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-858654f9db-knkd2"] Mar 17 09:23:45 crc kubenswrapper[4813]: I0317 09:23:45.968012 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-cf98fcc89-slrw5"] Mar 17 09:23:46 crc kubenswrapper[4813]: I0317 09:23:46.016441 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-687f57d79b-hdhpn"] Mar 17 09:23:46 crc kubenswrapper[4813]: I0317 09:23:46.481568 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-slrw5" event={"ID":"7ff1e0fd-1a34-4f70-97c0-9fbb47f0b52b","Type":"ContainerStarted","Data":"c7e9885fbcd51d84d74546122c78f95acaf3877f6d9680e84b21b970770a2ad1"} Mar 17 09:23:46 crc kubenswrapper[4813]: I0317 09:23:46.482882 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-hdhpn" event={"ID":"d949287d-f3e6-4ac8-8c16-a0d0652ae302","Type":"ContainerStarted","Data":"258adc80acb8766bcc85d1fe7b1bac6c0cc8a4d23a2a7d798c53f4e11b8d304c"} Mar 17 09:23:46 crc kubenswrapper[4813]: I0317 09:23:46.484050 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-knkd2" event={"ID":"08c8f504-aa0d-47b9-ab86-a4b440c6f446","Type":"ContainerStarted","Data":"16beb3187102424f69a6ea3023fb2c1c83707bc904fc979225ddce2f6e0baf20"} Mar 17 09:23:48 crc kubenswrapper[4813]: I0317 09:23:48.495061 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-cf98fcc89-slrw5" event={"ID":"7ff1e0fd-1a34-4f70-97c0-9fbb47f0b52b","Type":"ContainerStarted","Data":"a8bd85acc913ad4f01cda4920cf6814f6ba7b942d35c2cf35964d7636850c358"} Mar 17 09:23:48 crc kubenswrapper[4813]: I0317 09:23:48.524496 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-cf98fcc89-slrw5" podStartSLOduration=1.267087896 podStartE2EDuration="3.524472984s" podCreationTimestamp="2026-03-17 09:23:45 +0000 UTC" firstStartedPulling="2026-03-17 09:23:45.977402186 +0000 UTC m=+848.078205695" lastFinishedPulling="2026-03-17 09:23:48.234787284 +0000 UTC m=+850.335590783" observedRunningTime="2026-03-17 09:23:48.509640439 +0000 UTC m=+850.610443938" watchObservedRunningTime="2026-03-17 09:23:48.524472984 +0000 UTC m=+850.625276503" Mar 17 09:23:50 crc kubenswrapper[4813]: I0317 09:23:50.512322 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-687f57d79b-hdhpn" event={"ID":"d949287d-f3e6-4ac8-8c16-a0d0652ae302","Type":"ContainerStarted","Data":"17c454fda3d12527511de1e32cfd76ee046c76280119b6249053aa444bc28346"} Mar 17 09:23:50 crc kubenswrapper[4813]: I0317 09:23:50.512831 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-687f57d79b-hdhpn" Mar 17 09:23:50 crc kubenswrapper[4813]: I0317 09:23:50.515456 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-858654f9db-knkd2" event={"ID":"08c8f504-aa0d-47b9-ab86-a4b440c6f446","Type":"ContainerStarted","Data":"b2fc58136ab0b89a879cbcedf9d0374845ca3a9b8f37aa3620060b3e9811414c"} Mar 17 09:23:50 crc kubenswrapper[4813]: I0317 09:23:50.531633 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-687f57d79b-hdhpn" podStartSLOduration=1.835934033 podStartE2EDuration="5.53160812s" podCreationTimestamp="2026-03-17 09:23:45 +0000 UTC" firstStartedPulling="2026-03-17 09:23:46.026212975 +0000 UTC m=+848.127016494" lastFinishedPulling="2026-03-17 09:23:49.721887082 +0000 UTC m=+851.822690581" observedRunningTime="2026-03-17 09:23:50.529807003 +0000 UTC m=+852.630610532" watchObservedRunningTime="2026-03-17 09:23:50.53160812 +0000 UTC m=+852.632411639" Mar 17 09:23:50 crc kubenswrapper[4813]: I0317 09:23:50.553234 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-858654f9db-knkd2" podStartSLOduration=1.8064662889999998 podStartE2EDuration="5.553204437s" podCreationTimestamp="2026-03-17 09:23:45 +0000 UTC" firstStartedPulling="2026-03-17 09:23:45.931475346 +0000 UTC m=+848.032278845" lastFinishedPulling="2026-03-17 09:23:49.678213494 +0000 UTC m=+851.779016993" observedRunningTime="2026-03-17 09:23:50.549199722 +0000 UTC m=+852.650003261" watchObservedRunningTime="2026-03-17 09:23:50.553204437 +0000 UTC m=+852.654007966" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.486364 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lngs6"] Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.487595 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovn-controller" containerID="cri-o://7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7" gracePeriod=30 Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.487664 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="sbdb" containerID="cri-o://69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f" gracePeriod=30 Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.487788 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="kube-rbac-proxy-node" containerID="cri-o://0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded" gracePeriod=30 Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.487864 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovn-acl-logging" containerID="cri-o://791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed" gracePeriod=30 Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.487795 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd" gracePeriod=30 Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.487905 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="nbdb" containerID="cri-o://41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b" gracePeriod=30 Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.487947 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="northd" containerID="cri-o://27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923" gracePeriod=30 Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.534873 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovnkube-controller" containerID="cri-o://8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b" gracePeriod=30 Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.675975 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-687f57d79b-hdhpn" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.882656 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lngs6_a1730ca2-a1bb-4e9a-ba56-7edaded79f36/ovnkube-controller/3.log" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.886260 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lngs6_a1730ca2-a1bb-4e9a-ba56-7edaded79f36/ovn-acl-logging/0.log" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.886851 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lngs6_a1730ca2-a1bb-4e9a-ba56-7edaded79f36/ovn-controller/0.log" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.887368 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947215 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-jvl96"] Mar 17 09:23:55 crc kubenswrapper[4813]: E0317 09:23:55.947438 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovnkube-controller" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947453 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovnkube-controller" Mar 17 09:23:55 crc kubenswrapper[4813]: E0317 09:23:55.947465 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovnkube-controller" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947473 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovnkube-controller" Mar 17 09:23:55 crc kubenswrapper[4813]: E0317 09:23:55.947483 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovnkube-controller" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947491 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovnkube-controller" Mar 17 09:23:55 crc kubenswrapper[4813]: E0317 09:23:55.947501 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovn-acl-logging" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947509 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovn-acl-logging" Mar 17 09:23:55 crc kubenswrapper[4813]: E0317 09:23:55.947525 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="kubecfg-setup" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947532 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="kubecfg-setup" Mar 17 09:23:55 crc kubenswrapper[4813]: E0317 09:23:55.947541 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="northd" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947568 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="northd" Mar 17 09:23:55 crc kubenswrapper[4813]: E0317 09:23:55.947585 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="nbdb" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947593 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="nbdb" Mar 17 09:23:55 crc kubenswrapper[4813]: E0317 09:23:55.947625 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="kube-rbac-proxy-ovn-metrics" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947634 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="kube-rbac-proxy-ovn-metrics" Mar 17 09:23:55 crc kubenswrapper[4813]: E0317 09:23:55.947646 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovnkube-controller" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947654 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovnkube-controller" Mar 17 09:23:55 crc kubenswrapper[4813]: E0317 09:23:55.947665 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovn-controller" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947674 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovn-controller" Mar 17 09:23:55 crc kubenswrapper[4813]: E0317 09:23:55.947684 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovnkube-controller" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947692 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovnkube-controller" Mar 17 09:23:55 crc kubenswrapper[4813]: E0317 09:23:55.947702 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="sbdb" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947710 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="sbdb" Mar 17 09:23:55 crc kubenswrapper[4813]: E0317 09:23:55.947722 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="kube-rbac-proxy-node" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947729 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="kube-rbac-proxy-node" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947834 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovnkube-controller" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947845 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="sbdb" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947854 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovn-controller" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947868 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovn-acl-logging" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947879 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovnkube-controller" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947892 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovnkube-controller" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947904 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="nbdb" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947917 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="kube-rbac-proxy-node" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947930 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="northd" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.947940 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="kube-rbac-proxy-ovn-metrics" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.948207 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovnkube-controller" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.948223 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerName="ovnkube-controller" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.950219 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.984672 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j4jw5\" (UniqueName: \"kubernetes.io/projected/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-kube-api-access-j4jw5\") pod \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.984709 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-run-netns\") pod \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.984730 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-systemd-units\") pod \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.984756 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-ovn-node-metrics-cert\") pod \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.984778 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-var-lib-openvswitch\") pod \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.984808 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-ovnkube-script-lib\") pod \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.984827 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-env-overrides\") pod \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.984851 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-run-ovn-kubernetes\") pod \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.984884 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-kubelet\") pod \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.984875 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "a1730ca2-a1bb-4e9a-ba56-7edaded79f36" (UID: "a1730ca2-a1bb-4e9a-ba56-7edaded79f36"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.984942 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "a1730ca2-a1bb-4e9a-ba56-7edaded79f36" (UID: "a1730ca2-a1bb-4e9a-ba56-7edaded79f36"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.985081 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "a1730ca2-a1bb-4e9a-ba56-7edaded79f36" (UID: "a1730ca2-a1bb-4e9a-ba56-7edaded79f36"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.985519 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "a1730ca2-a1bb-4e9a-ba56-7edaded79f36" (UID: "a1730ca2-a1bb-4e9a-ba56-7edaded79f36"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.985540 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "a1730ca2-a1bb-4e9a-ba56-7edaded79f36" (UID: "a1730ca2-a1bb-4e9a-ba56-7edaded79f36"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.985620 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-cni-bin\") pod \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.985646 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-var-lib-cni-networks-ovn-kubernetes\") pod \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.985673 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-node-log\") pod \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.985639 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "a1730ca2-a1bb-4e9a-ba56-7edaded79f36" (UID: "a1730ca2-a1bb-4e9a-ba56-7edaded79f36"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.985713 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "a1730ca2-a1bb-4e9a-ba56-7edaded79f36" (UID: "a1730ca2-a1bb-4e9a-ba56-7edaded79f36"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.985742 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-slash\") pod \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.985763 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-etc-openvswitch\") pod \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.985675 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "a1730ca2-a1bb-4e9a-ba56-7edaded79f36" (UID: "a1730ca2-a1bb-4e9a-ba56-7edaded79f36"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.985763 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-node-log" (OuterVolumeSpecName: "node-log") pod "a1730ca2-a1bb-4e9a-ba56-7edaded79f36" (UID: "a1730ca2-a1bb-4e9a-ba56-7edaded79f36"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.985809 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-slash" (OuterVolumeSpecName: "host-slash") pod "a1730ca2-a1bb-4e9a-ba56-7edaded79f36" (UID: "a1730ca2-a1bb-4e9a-ba56-7edaded79f36"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.985835 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-ovnkube-config\") pod \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.985889 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "a1730ca2-a1bb-4e9a-ba56-7edaded79f36" (UID: "a1730ca2-a1bb-4e9a-ba56-7edaded79f36"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.985915 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-run-systemd\") pod \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.985937 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-run-ovn\") pod \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.986071 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "a1730ca2-a1bb-4e9a-ba56-7edaded79f36" (UID: "a1730ca2-a1bb-4e9a-ba56-7edaded79f36"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.984840 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "a1730ca2-a1bb-4e9a-ba56-7edaded79f36" (UID: "a1730ca2-a1bb-4e9a-ba56-7edaded79f36"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.986372 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "a1730ca2-a1bb-4e9a-ba56-7edaded79f36" (UID: "a1730ca2-a1bb-4e9a-ba56-7edaded79f36"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.986464 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-log-socket\") pod \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.986495 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-run-openvswitch\") pod \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.986522 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-cni-netd\") pod \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\" (UID: \"a1730ca2-a1bb-4e9a-ba56-7edaded79f36\") " Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.986647 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "a1730ca2-a1bb-4e9a-ba56-7edaded79f36" (UID: "a1730ca2-a1bb-4e9a-ba56-7edaded79f36"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.986683 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "a1730ca2-a1bb-4e9a-ba56-7edaded79f36" (UID: "a1730ca2-a1bb-4e9a-ba56-7edaded79f36"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.986924 4813 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.986948 4813 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.986964 4813 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.986978 4813 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.986991 4813 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.987005 4813 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.987019 4813 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.987037 4813 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.987051 4813 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.987065 4813 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.987080 4813 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.987093 4813 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.987107 4813 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.987123 4813 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-node-log\") on node \"crc\" DevicePath \"\"" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.987138 4813 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-host-slash\") on node \"crc\" DevicePath \"\"" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.987152 4813 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.987198 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-log-socket" (OuterVolumeSpecName: "log-socket") pod "a1730ca2-a1bb-4e9a-ba56-7edaded79f36" (UID: "a1730ca2-a1bb-4e9a-ba56-7edaded79f36"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.996412 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-kube-api-access-j4jw5" (OuterVolumeSpecName: "kube-api-access-j4jw5") pod "a1730ca2-a1bb-4e9a-ba56-7edaded79f36" (UID: "a1730ca2-a1bb-4e9a-ba56-7edaded79f36"). InnerVolumeSpecName "kube-api-access-j4jw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:23:55 crc kubenswrapper[4813]: I0317 09:23:55.997183 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "a1730ca2-a1bb-4e9a-ba56-7edaded79f36" (UID: "a1730ca2-a1bb-4e9a-ba56-7edaded79f36"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.012931 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "a1730ca2-a1bb-4e9a-ba56-7edaded79f36" (UID: "a1730ca2-a1bb-4e9a-ba56-7edaded79f36"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.088345 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-systemd-units\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.088425 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-node-log\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.088460 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-host-run-ovn-kubernetes\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.088499 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-run-openvswitch\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.088538 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-log-socket\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.088577 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-host-kubelet\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.088689 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-etc-openvswitch\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.088740 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-var-lib-openvswitch\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.088775 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-run-systemd\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.088808 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dfaad66f-3e5b-45dc-9391-c6249f85cdea-ovn-node-metrics-cert\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.088841 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-run-ovn\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.088866 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dfaad66f-3e5b-45dc-9391-c6249f85cdea-ovnkube-config\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.088927 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr92f\" (UniqueName: \"kubernetes.io/projected/dfaad66f-3e5b-45dc-9391-c6249f85cdea-kube-api-access-hr92f\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.088955 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-host-cni-netd\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.088977 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-host-slash\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.089001 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.089094 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/dfaad66f-3e5b-45dc-9391-c6249f85cdea-ovnkube-script-lib\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.089133 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-host-cni-bin\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.089162 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dfaad66f-3e5b-45dc-9391-c6249f85cdea-env-overrides\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.089219 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-host-run-netns\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.089290 4813 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.089320 4813 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-log-socket\") on node \"crc\" DevicePath \"\"" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.089349 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j4jw5\" (UniqueName: \"kubernetes.io/projected/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-kube-api-access-j4jw5\") on node \"crc\" DevicePath \"\"" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.089377 4813 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/a1730ca2-a1bb-4e9a-ba56-7edaded79f36-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191024 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-run-ovn\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191091 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dfaad66f-3e5b-45dc-9391-c6249f85cdea-ovnkube-config\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191122 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-run-ovn\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191132 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr92f\" (UniqueName: \"kubernetes.io/projected/dfaad66f-3e5b-45dc-9391-c6249f85cdea-kube-api-access-hr92f\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191212 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-host-cni-netd\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191246 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-host-slash\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191284 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191315 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/dfaad66f-3e5b-45dc-9391-c6249f85cdea-ovnkube-script-lib\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191343 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-host-cni-bin\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191364 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dfaad66f-3e5b-45dc-9391-c6249f85cdea-env-overrides\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191390 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-host-run-netns\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191414 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-systemd-units\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191433 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-node-log\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191455 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-host-run-ovn-kubernetes\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191474 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-run-openvswitch\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191495 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-log-socket\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191519 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-host-kubelet\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191539 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-etc-openvswitch\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191562 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-var-lib-openvswitch\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191585 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-run-systemd\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191662 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dfaad66f-3e5b-45dc-9391-c6249f85cdea-ovn-node-metrics-cert\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.191722 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-host-run-ovn-kubernetes\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.192068 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-host-cni-bin\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.192096 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-run-systemd\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.192090 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-host-kubelet\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.192118 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-host-run-netns\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.192125 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-var-lib-openvswitch\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.192170 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-systemd-units\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.192200 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-node-log\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.192226 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-etc-openvswitch\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.192257 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-run-openvswitch\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.192283 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-log-socket\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.192288 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/dfaad66f-3e5b-45dc-9391-c6249f85cdea-env-overrides\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.192322 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-host-cni-netd\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.192335 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-host-slash\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.192351 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/dfaad66f-3e5b-45dc-9391-c6249f85cdea-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.192594 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/dfaad66f-3e5b-45dc-9391-c6249f85cdea-ovnkube-config\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.193094 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/dfaad66f-3e5b-45dc-9391-c6249f85cdea-ovnkube-script-lib\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.195984 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/dfaad66f-3e5b-45dc-9391-c6249f85cdea-ovn-node-metrics-cert\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.218175 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr92f\" (UniqueName: \"kubernetes.io/projected/dfaad66f-3e5b-45dc-9391-c6249f85cdea-kube-api-access-hr92f\") pod \"ovnkube-node-jvl96\" (UID: \"dfaad66f-3e5b-45dc-9391-c6249f85cdea\") " pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.284158 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.559302 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5jjhb_c1246d4d-93d5-4a97-bef8-1ed881e1a217/kube-multus/2.log" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.560924 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5jjhb_c1246d4d-93d5-4a97-bef8-1ed881e1a217/kube-multus/1.log" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.561007 4813 generic.go:334] "Generic (PLEG): container finished" podID="c1246d4d-93d5-4a97-bef8-1ed881e1a217" containerID="b72d7d564b1340f9238d2a5b0d3bda56b2ecbeb27116e4100944814dcef2fb8e" exitCode=2 Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.561108 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5jjhb" event={"ID":"c1246d4d-93d5-4a97-bef8-1ed881e1a217","Type":"ContainerDied","Data":"b72d7d564b1340f9238d2a5b0d3bda56b2ecbeb27116e4100944814dcef2fb8e"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.561199 4813 scope.go:117] "RemoveContainer" containerID="3cf6e87ac86b646f6e7e520f8dccf5788696ddef35970778f9490f817a04385a" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.561904 4813 scope.go:117] "RemoveContainer" containerID="b72d7d564b1340f9238d2a5b0d3bda56b2ecbeb27116e4100944814dcef2fb8e" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.564973 4813 generic.go:334] "Generic (PLEG): container finished" podID="dfaad66f-3e5b-45dc-9391-c6249f85cdea" containerID="0dbeb803b40f0070bfd84ade6129b52f6e2b596ff15b8210ba4801101d4c29b5" exitCode=0 Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.565080 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" event={"ID":"dfaad66f-3e5b-45dc-9391-c6249f85cdea","Type":"ContainerDied","Data":"0dbeb803b40f0070bfd84ade6129b52f6e2b596ff15b8210ba4801101d4c29b5"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.565152 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" event={"ID":"dfaad66f-3e5b-45dc-9391-c6249f85cdea","Type":"ContainerStarted","Data":"e8a9b3d627c1f8903cb3615070eaad307b4f75b3798ff07929e31b2c6734512d"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.580207 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lngs6_a1730ca2-a1bb-4e9a-ba56-7edaded79f36/ovnkube-controller/3.log" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.591511 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lngs6_a1730ca2-a1bb-4e9a-ba56-7edaded79f36/ovn-acl-logging/0.log" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.593586 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lngs6_a1730ca2-a1bb-4e9a-ba56-7edaded79f36/ovn-controller/0.log" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.595222 4813 generic.go:334] "Generic (PLEG): container finished" podID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerID="8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b" exitCode=0 Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.595357 4813 generic.go:334] "Generic (PLEG): container finished" podID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerID="69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f" exitCode=0 Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.595456 4813 generic.go:334] "Generic (PLEG): container finished" podID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerID="41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b" exitCode=0 Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.595527 4813 generic.go:334] "Generic (PLEG): container finished" podID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerID="27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923" exitCode=0 Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.595554 4813 generic.go:334] "Generic (PLEG): container finished" podID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerID="dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd" exitCode=0 Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.595567 4813 generic.go:334] "Generic (PLEG): container finished" podID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerID="0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded" exitCode=0 Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.595733 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596273 4813 generic.go:334] "Generic (PLEG): container finished" podID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerID="791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed" exitCode=143 Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596315 4813 generic.go:334] "Generic (PLEG): container finished" podID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" containerID="7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7" exitCode=143 Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.595258 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerDied","Data":"8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596509 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerDied","Data":"69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596585 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerDied","Data":"41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596624 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerDied","Data":"27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596657 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerDied","Data":"dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596672 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerDied","Data":"0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596687 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596702 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596709 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596717 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596748 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596762 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596770 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596777 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596784 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596790 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596800 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerDied","Data":"791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596816 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596824 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596830 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596837 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596844 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596851 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596858 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596866 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596878 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596887 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596896 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerDied","Data":"7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596906 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596914 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596922 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596928 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596936 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596942 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596948 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.596955 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.597197 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.597205 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.597224 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lngs6" event={"ID":"a1730ca2-a1bb-4e9a-ba56-7edaded79f36","Type":"ContainerDied","Data":"4e31eab6db6123fdbd4bd448a5938c83e568c948e29eec8e6cb4a07dc82b2f09"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.597240 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.597248 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.597255 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.597262 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.597269 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.597275 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.597282 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.597288 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.597295 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.597301 4813 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5"} Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.630694 4813 scope.go:117] "RemoveContainer" containerID="8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.656159 4813 scope.go:117] "RemoveContainer" containerID="09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.681123 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lngs6"] Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.686187 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lngs6"] Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.702040 4813 scope.go:117] "RemoveContainer" containerID="69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.738910 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1730ca2-a1bb-4e9a-ba56-7edaded79f36" path="/var/lib/kubelet/pods/a1730ca2-a1bb-4e9a-ba56-7edaded79f36/volumes" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.747046 4813 scope.go:117] "RemoveContainer" containerID="41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.765893 4813 scope.go:117] "RemoveContainer" containerID="27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.787437 4813 scope.go:117] "RemoveContainer" containerID="dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.803339 4813 scope.go:117] "RemoveContainer" containerID="0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.829448 4813 scope.go:117] "RemoveContainer" containerID="791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.860877 4813 scope.go:117] "RemoveContainer" containerID="7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.878349 4813 scope.go:117] "RemoveContainer" containerID="06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.909016 4813 scope.go:117] "RemoveContainer" containerID="8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b" Mar 17 09:23:56 crc kubenswrapper[4813]: E0317 09:23:56.909427 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b\": container with ID starting with 8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b not found: ID does not exist" containerID="8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.909463 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b"} err="failed to get container status \"8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b\": rpc error: code = NotFound desc = could not find container \"8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b\": container with ID starting with 8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.909490 4813 scope.go:117] "RemoveContainer" containerID="09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b" Mar 17 09:23:56 crc kubenswrapper[4813]: E0317 09:23:56.909904 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b\": container with ID starting with 09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b not found: ID does not exist" containerID="09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.909935 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b"} err="failed to get container status \"09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b\": rpc error: code = NotFound desc = could not find container \"09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b\": container with ID starting with 09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.909952 4813 scope.go:117] "RemoveContainer" containerID="69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f" Mar 17 09:23:56 crc kubenswrapper[4813]: E0317 09:23:56.910336 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\": container with ID starting with 69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f not found: ID does not exist" containerID="69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.910410 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f"} err="failed to get container status \"69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\": rpc error: code = NotFound desc = could not find container \"69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\": container with ID starting with 69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.910466 4813 scope.go:117] "RemoveContainer" containerID="41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b" Mar 17 09:23:56 crc kubenswrapper[4813]: E0317 09:23:56.911247 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\": container with ID starting with 41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b not found: ID does not exist" containerID="41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.911313 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b"} err="failed to get container status \"41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\": rpc error: code = NotFound desc = could not find container \"41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\": container with ID starting with 41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.911358 4813 scope.go:117] "RemoveContainer" containerID="27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923" Mar 17 09:23:56 crc kubenswrapper[4813]: E0317 09:23:56.911792 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\": container with ID starting with 27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923 not found: ID does not exist" containerID="27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.911847 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923"} err="failed to get container status \"27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\": rpc error: code = NotFound desc = could not find container \"27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\": container with ID starting with 27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923 not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.911991 4813 scope.go:117] "RemoveContainer" containerID="dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd" Mar 17 09:23:56 crc kubenswrapper[4813]: E0317 09:23:56.912336 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\": container with ID starting with dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd not found: ID does not exist" containerID="dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.912364 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd"} err="failed to get container status \"dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\": rpc error: code = NotFound desc = could not find container \"dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\": container with ID starting with dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.912384 4813 scope.go:117] "RemoveContainer" containerID="0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded" Mar 17 09:23:56 crc kubenswrapper[4813]: E0317 09:23:56.912746 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\": container with ID starting with 0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded not found: ID does not exist" containerID="0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.912770 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded"} err="failed to get container status \"0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\": rpc error: code = NotFound desc = could not find container \"0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\": container with ID starting with 0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.912787 4813 scope.go:117] "RemoveContainer" containerID="791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed" Mar 17 09:23:56 crc kubenswrapper[4813]: E0317 09:23:56.913237 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\": container with ID starting with 791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed not found: ID does not exist" containerID="791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.913258 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed"} err="failed to get container status \"791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\": rpc error: code = NotFound desc = could not find container \"791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\": container with ID starting with 791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.913275 4813 scope.go:117] "RemoveContainer" containerID="7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7" Mar 17 09:23:56 crc kubenswrapper[4813]: E0317 09:23:56.913486 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\": container with ID starting with 7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7 not found: ID does not exist" containerID="7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.913505 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7"} err="failed to get container status \"7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\": rpc error: code = NotFound desc = could not find container \"7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\": container with ID starting with 7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7 not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.913520 4813 scope.go:117] "RemoveContainer" containerID="06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5" Mar 17 09:23:56 crc kubenswrapper[4813]: E0317 09:23:56.913725 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\": container with ID starting with 06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5 not found: ID does not exist" containerID="06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.913748 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5"} err="failed to get container status \"06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\": rpc error: code = NotFound desc = could not find container \"06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\": container with ID starting with 06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5 not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.913767 4813 scope.go:117] "RemoveContainer" containerID="8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.914129 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b"} err="failed to get container status \"8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b\": rpc error: code = NotFound desc = could not find container \"8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b\": container with ID starting with 8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.914150 4813 scope.go:117] "RemoveContainer" containerID="09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.914664 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b"} err="failed to get container status \"09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b\": rpc error: code = NotFound desc = could not find container \"09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b\": container with ID starting with 09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.914683 4813 scope.go:117] "RemoveContainer" containerID="69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.915289 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f"} err="failed to get container status \"69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\": rpc error: code = NotFound desc = could not find container \"69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\": container with ID starting with 69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.915413 4813 scope.go:117] "RemoveContainer" containerID="41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.919922 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b"} err="failed to get container status \"41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\": rpc error: code = NotFound desc = could not find container \"41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\": container with ID starting with 41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.919978 4813 scope.go:117] "RemoveContainer" containerID="27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.922899 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923"} err="failed to get container status \"27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\": rpc error: code = NotFound desc = could not find container \"27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\": container with ID starting with 27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923 not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.922954 4813 scope.go:117] "RemoveContainer" containerID="dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.923585 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd"} err="failed to get container status \"dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\": rpc error: code = NotFound desc = could not find container \"dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\": container with ID starting with dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.923675 4813 scope.go:117] "RemoveContainer" containerID="0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.924196 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded"} err="failed to get container status \"0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\": rpc error: code = NotFound desc = could not find container \"0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\": container with ID starting with 0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.924244 4813 scope.go:117] "RemoveContainer" containerID="791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.924654 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed"} err="failed to get container status \"791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\": rpc error: code = NotFound desc = could not find container \"791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\": container with ID starting with 791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.924728 4813 scope.go:117] "RemoveContainer" containerID="7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.925336 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7"} err="failed to get container status \"7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\": rpc error: code = NotFound desc = could not find container \"7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\": container with ID starting with 7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7 not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.925382 4813 scope.go:117] "RemoveContainer" containerID="06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.930990 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5"} err="failed to get container status \"06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\": rpc error: code = NotFound desc = could not find container \"06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\": container with ID starting with 06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5 not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.931068 4813 scope.go:117] "RemoveContainer" containerID="8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.931836 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b"} err="failed to get container status \"8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b\": rpc error: code = NotFound desc = could not find container \"8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b\": container with ID starting with 8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.931916 4813 scope.go:117] "RemoveContainer" containerID="09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.932629 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b"} err="failed to get container status \"09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b\": rpc error: code = NotFound desc = could not find container \"09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b\": container with ID starting with 09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.932679 4813 scope.go:117] "RemoveContainer" containerID="69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.933183 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f"} err="failed to get container status \"69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\": rpc error: code = NotFound desc = could not find container \"69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\": container with ID starting with 69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.933205 4813 scope.go:117] "RemoveContainer" containerID="41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.933709 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b"} err="failed to get container status \"41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\": rpc error: code = NotFound desc = could not find container \"41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\": container with ID starting with 41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.933733 4813 scope.go:117] "RemoveContainer" containerID="27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.934288 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923"} err="failed to get container status \"27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\": rpc error: code = NotFound desc = could not find container \"27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\": container with ID starting with 27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923 not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.934336 4813 scope.go:117] "RemoveContainer" containerID="dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.934784 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd"} err="failed to get container status \"dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\": rpc error: code = NotFound desc = could not find container \"dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\": container with ID starting with dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.934809 4813 scope.go:117] "RemoveContainer" containerID="0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.935468 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded"} err="failed to get container status \"0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\": rpc error: code = NotFound desc = could not find container \"0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\": container with ID starting with 0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.935528 4813 scope.go:117] "RemoveContainer" containerID="791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.936339 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed"} err="failed to get container status \"791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\": rpc error: code = NotFound desc = could not find container \"791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\": container with ID starting with 791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.936415 4813 scope.go:117] "RemoveContainer" containerID="7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.936879 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7"} err="failed to get container status \"7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\": rpc error: code = NotFound desc = could not find container \"7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\": container with ID starting with 7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7 not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.936933 4813 scope.go:117] "RemoveContainer" containerID="06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.937595 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5"} err="failed to get container status \"06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\": rpc error: code = NotFound desc = could not find container \"06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\": container with ID starting with 06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5 not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.937705 4813 scope.go:117] "RemoveContainer" containerID="8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.938107 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b"} err="failed to get container status \"8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b\": rpc error: code = NotFound desc = could not find container \"8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b\": container with ID starting with 8a9b804dccf7e2cdc893dabeea11047c1797cffeec0abd3f4f0978b70fb6f98b not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.938173 4813 scope.go:117] "RemoveContainer" containerID="09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.939741 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b"} err="failed to get container status \"09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b\": rpc error: code = NotFound desc = could not find container \"09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b\": container with ID starting with 09c9d74398e15108a537b377e4eef423ff4476fc1744759b8d9b1497e9fa4c7b not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.939830 4813 scope.go:117] "RemoveContainer" containerID="69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.941645 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f"} err="failed to get container status \"69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\": rpc error: code = NotFound desc = could not find container \"69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f\": container with ID starting with 69d9e6f0d6c2ea016d40e91d456ab3bea590e9adaa45b982a7a58ec2c9332f4f not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.941718 4813 scope.go:117] "RemoveContainer" containerID="41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.942477 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b"} err="failed to get container status \"41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\": rpc error: code = NotFound desc = could not find container \"41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b\": container with ID starting with 41cc994ed8d4c6c91958e19108e55facdf4ff0d927bd2e1afa9ed4f61ed6ff2b not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.942545 4813 scope.go:117] "RemoveContainer" containerID="27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.942967 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923"} err="failed to get container status \"27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\": rpc error: code = NotFound desc = could not find container \"27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923\": container with ID starting with 27311cb1c43f03a6f49e9badfe3fa063f1990b24089d71710da469367532a923 not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.943060 4813 scope.go:117] "RemoveContainer" containerID="dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.943758 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd"} err="failed to get container status \"dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\": rpc error: code = NotFound desc = could not find container \"dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd\": container with ID starting with dd7568f418c2335774aa344ea4262ff51f49278aca26dfb64c0dd5eaa747c9dd not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.943847 4813 scope.go:117] "RemoveContainer" containerID="0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.944454 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded"} err="failed to get container status \"0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\": rpc error: code = NotFound desc = could not find container \"0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded\": container with ID starting with 0298f881ad44f7b11bbefda5be7ff988cde684b27f31e18448015b6728da9ded not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.944532 4813 scope.go:117] "RemoveContainer" containerID="791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.944849 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed"} err="failed to get container status \"791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\": rpc error: code = NotFound desc = could not find container \"791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed\": container with ID starting with 791defcbba553273c33526b29147a0bfffadda9924c964a9cf3219ccbcefe0ed not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.944889 4813 scope.go:117] "RemoveContainer" containerID="7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.945431 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7"} err="failed to get container status \"7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\": rpc error: code = NotFound desc = could not find container \"7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7\": container with ID starting with 7d7b70120c23501a1f43e191374c79a02f3da97cd35bd1fd43410e6ca1714ad7 not found: ID does not exist" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.945488 4813 scope.go:117] "RemoveContainer" containerID="06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5" Mar 17 09:23:56 crc kubenswrapper[4813]: I0317 09:23:56.945904 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5"} err="failed to get container status \"06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\": rpc error: code = NotFound desc = could not find container \"06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5\": container with ID starting with 06c6cab19ef325a36be1ac8ccfe22cbb3092089ecbae6c21711f0c24449e72c5 not found: ID does not exist" Mar 17 09:23:57 crc kubenswrapper[4813]: I0317 09:23:57.127809 4813 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 17 09:23:57 crc kubenswrapper[4813]: I0317 09:23:57.615466 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-5jjhb_c1246d4d-93d5-4a97-bef8-1ed881e1a217/kube-multus/2.log" Mar 17 09:23:57 crc kubenswrapper[4813]: I0317 09:23:57.616015 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-5jjhb" event={"ID":"c1246d4d-93d5-4a97-bef8-1ed881e1a217","Type":"ContainerStarted","Data":"9c061334a8602b807bcd59c832d10034ab80cb20f98b54eb2fdc7efcc49db5ef"} Mar 17 09:23:57 crc kubenswrapper[4813]: I0317 09:23:57.624335 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" event={"ID":"dfaad66f-3e5b-45dc-9391-c6249f85cdea","Type":"ContainerStarted","Data":"4975b3827a1a29224ac83cbf0f12c7a8a40a3955d75e14403ccb8d8e80a8666e"} Mar 17 09:23:57 crc kubenswrapper[4813]: I0317 09:23:57.624383 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" event={"ID":"dfaad66f-3e5b-45dc-9391-c6249f85cdea","Type":"ContainerStarted","Data":"cb4abdcfb026e5f65d4b2e8018bd7df24910ab1c9032abefebc3c357a1db3c2b"} Mar 17 09:23:57 crc kubenswrapper[4813]: I0317 09:23:57.624399 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" event={"ID":"dfaad66f-3e5b-45dc-9391-c6249f85cdea","Type":"ContainerStarted","Data":"7a5e91ef36c0b72231eb52204d81baad9ff051fdd51fc00af2ff9379cedf1543"} Mar 17 09:23:57 crc kubenswrapper[4813]: I0317 09:23:57.624410 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" event={"ID":"dfaad66f-3e5b-45dc-9391-c6249f85cdea","Type":"ContainerStarted","Data":"8b66b342d325939e52ef9592dd04a1d4e7527a953afb2d71a8118673930cd594"} Mar 17 09:23:57 crc kubenswrapper[4813]: I0317 09:23:57.624420 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" event={"ID":"dfaad66f-3e5b-45dc-9391-c6249f85cdea","Type":"ContainerStarted","Data":"0ab10e4c5fa8e1da1297cd11773c3a4d3c462341cf7254ebca3f18215293dbd1"} Mar 17 09:23:57 crc kubenswrapper[4813]: I0317 09:23:57.624429 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" event={"ID":"dfaad66f-3e5b-45dc-9391-c6249f85cdea","Type":"ContainerStarted","Data":"990b7f79b9150e84b5be52730bcc0c3a5f799321c15da7555122f2f9afddf7bd"} Mar 17 09:23:59 crc kubenswrapper[4813]: I0317 09:23:59.641639 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" event={"ID":"dfaad66f-3e5b-45dc-9391-c6249f85cdea","Type":"ContainerStarted","Data":"88ec1a7cb613474ce5ae340f3c58d24d1bbc180a3907f782d3308e555b7e159c"} Mar 17 09:24:00 crc kubenswrapper[4813]: I0317 09:24:00.145258 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562324-zfp5g"] Mar 17 09:24:00 crc kubenswrapper[4813]: I0317 09:24:00.145905 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562324-zfp5g" Mar 17 09:24:00 crc kubenswrapper[4813]: I0317 09:24:00.148373 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:24:00 crc kubenswrapper[4813]: I0317 09:24:00.148793 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:24:00 crc kubenswrapper[4813]: I0317 09:24:00.149122 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:24:00 crc kubenswrapper[4813]: I0317 09:24:00.151486 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzx8g\" (UniqueName: \"kubernetes.io/projected/9a6501c3-1a60-4c0d-b190-5a8c4335b527-kube-api-access-fzx8g\") pod \"auto-csr-approver-29562324-zfp5g\" (UID: \"9a6501c3-1a60-4c0d-b190-5a8c4335b527\") " pod="openshift-infra/auto-csr-approver-29562324-zfp5g" Mar 17 09:24:00 crc kubenswrapper[4813]: I0317 09:24:00.252541 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzx8g\" (UniqueName: \"kubernetes.io/projected/9a6501c3-1a60-4c0d-b190-5a8c4335b527-kube-api-access-fzx8g\") pod \"auto-csr-approver-29562324-zfp5g\" (UID: \"9a6501c3-1a60-4c0d-b190-5a8c4335b527\") " pod="openshift-infra/auto-csr-approver-29562324-zfp5g" Mar 17 09:24:00 crc kubenswrapper[4813]: I0317 09:24:00.272201 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzx8g\" (UniqueName: \"kubernetes.io/projected/9a6501c3-1a60-4c0d-b190-5a8c4335b527-kube-api-access-fzx8g\") pod \"auto-csr-approver-29562324-zfp5g\" (UID: \"9a6501c3-1a60-4c0d-b190-5a8c4335b527\") " pod="openshift-infra/auto-csr-approver-29562324-zfp5g" Mar 17 09:24:00 crc kubenswrapper[4813]: I0317 09:24:00.470751 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562324-zfp5g" Mar 17 09:24:00 crc kubenswrapper[4813]: E0317 09:24:00.506198 4813 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29562324-zfp5g_openshift-infra_9a6501c3-1a60-4c0d-b190-5a8c4335b527_0(dbbcfa5a6933b497da5a1ccc198a7cb305be8b8f7a1ea06c9decf870f1fe7402): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:24:00 crc kubenswrapper[4813]: E0317 09:24:00.506327 4813 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29562324-zfp5g_openshift-infra_9a6501c3-1a60-4c0d-b190-5a8c4335b527_0(dbbcfa5a6933b497da5a1ccc198a7cb305be8b8f7a1ea06c9decf870f1fe7402): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29562324-zfp5g" Mar 17 09:24:00 crc kubenswrapper[4813]: E0317 09:24:00.506365 4813 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29562324-zfp5g_openshift-infra_9a6501c3-1a60-4c0d-b190-5a8c4335b527_0(dbbcfa5a6933b497da5a1ccc198a7cb305be8b8f7a1ea06c9decf870f1fe7402): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29562324-zfp5g" Mar 17 09:24:00 crc kubenswrapper[4813]: E0317 09:24:00.506437 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"auto-csr-approver-29562324-zfp5g_openshift-infra(9a6501c3-1a60-4c0d-b190-5a8c4335b527)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"auto-csr-approver-29562324-zfp5g_openshift-infra(9a6501c3-1a60-4c0d-b190-5a8c4335b527)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29562324-zfp5g_openshift-infra_9a6501c3-1a60-4c0d-b190-5a8c4335b527_0(dbbcfa5a6933b497da5a1ccc198a7cb305be8b8f7a1ea06c9decf870f1fe7402): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-infra/auto-csr-approver-29562324-zfp5g" podUID="9a6501c3-1a60-4c0d-b190-5a8c4335b527" Mar 17 09:24:02 crc kubenswrapper[4813]: I0317 09:24:02.666380 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" event={"ID":"dfaad66f-3e5b-45dc-9391-c6249f85cdea","Type":"ContainerStarted","Data":"035c11de9b8afd6bcb9b627691c7abd9bcdc0c764ec094e586ef3b303961d2fb"} Mar 17 09:24:02 crc kubenswrapper[4813]: I0317 09:24:02.667055 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:24:02 crc kubenswrapper[4813]: I0317 09:24:02.707835 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" podStartSLOduration=7.707817327 podStartE2EDuration="7.707817327s" podCreationTimestamp="2026-03-17 09:23:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:24:02.705225775 +0000 UTC m=+864.806029284" watchObservedRunningTime="2026-03-17 09:24:02.707817327 +0000 UTC m=+864.808620826" Mar 17 09:24:02 crc kubenswrapper[4813]: I0317 09:24:02.747869 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:24:03 crc kubenswrapper[4813]: I0317 09:24:03.671329 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:24:03 crc kubenswrapper[4813]: I0317 09:24:03.671376 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:24:03 crc kubenswrapper[4813]: I0317 09:24:03.731984 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:24:03 crc kubenswrapper[4813]: I0317 09:24:03.977057 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562324-zfp5g"] Mar 17 09:24:03 crc kubenswrapper[4813]: I0317 09:24:03.977189 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562324-zfp5g" Mar 17 09:24:03 crc kubenswrapper[4813]: I0317 09:24:03.977615 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562324-zfp5g" Mar 17 09:24:04 crc kubenswrapper[4813]: E0317 09:24:04.024008 4813 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29562324-zfp5g_openshift-infra_9a6501c3-1a60-4c0d-b190-5a8c4335b527_0(185e9241300a6cb259808af109d78ae62a29187f3856eb1eb8f6147588720ecd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 17 09:24:04 crc kubenswrapper[4813]: E0317 09:24:04.024090 4813 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29562324-zfp5g_openshift-infra_9a6501c3-1a60-4c0d-b190-5a8c4335b527_0(185e9241300a6cb259808af109d78ae62a29187f3856eb1eb8f6147588720ecd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29562324-zfp5g" Mar 17 09:24:04 crc kubenswrapper[4813]: E0317 09:24:04.024124 4813 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29562324-zfp5g_openshift-infra_9a6501c3-1a60-4c0d-b190-5a8c4335b527_0(185e9241300a6cb259808af109d78ae62a29187f3856eb1eb8f6147588720ecd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-infra/auto-csr-approver-29562324-zfp5g" Mar 17 09:24:04 crc kubenswrapper[4813]: E0317 09:24:04.024187 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"auto-csr-approver-29562324-zfp5g_openshift-infra(9a6501c3-1a60-4c0d-b190-5a8c4335b527)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"auto-csr-approver-29562324-zfp5g_openshift-infra(9a6501c3-1a60-4c0d-b190-5a8c4335b527)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_auto-csr-approver-29562324-zfp5g_openshift-infra_9a6501c3-1a60-4c0d-b190-5a8c4335b527_0(185e9241300a6cb259808af109d78ae62a29187f3856eb1eb8f6147588720ecd): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-infra/auto-csr-approver-29562324-zfp5g" podUID="9a6501c3-1a60-4c0d-b190-5a8c4335b527" Mar 17 09:24:14 crc kubenswrapper[4813]: I0317 09:24:14.113969 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:24:14 crc kubenswrapper[4813]: I0317 09:24:14.114589 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:24:14 crc kubenswrapper[4813]: I0317 09:24:14.114688 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:24:14 crc kubenswrapper[4813]: I0317 09:24:14.115542 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b67bb2399345967e5d98c3e93bfb13d1f15014e04cd83984dc098f8a87b2a52c"} pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 09:24:14 crc kubenswrapper[4813]: I0317 09:24:14.115654 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" containerID="cri-o://b67bb2399345967e5d98c3e93bfb13d1f15014e04cd83984dc098f8a87b2a52c" gracePeriod=600 Mar 17 09:24:14 crc kubenswrapper[4813]: I0317 09:24:14.730645 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562324-zfp5g" Mar 17 09:24:14 crc kubenswrapper[4813]: I0317 09:24:14.731471 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562324-zfp5g" Mar 17 09:24:15 crc kubenswrapper[4813]: I0317 09:24:15.098509 4813 generic.go:334] "Generic (PLEG): container finished" podID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerID="b67bb2399345967e5d98c3e93bfb13d1f15014e04cd83984dc098f8a87b2a52c" exitCode=0 Mar 17 09:24:15 crc kubenswrapper[4813]: I0317 09:24:15.098572 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerDied","Data":"b67bb2399345967e5d98c3e93bfb13d1f15014e04cd83984dc098f8a87b2a52c"} Mar 17 09:24:15 crc kubenswrapper[4813]: I0317 09:24:15.098922 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerStarted","Data":"c178911cb0823cd7d042791f6122482441f05cc74687451c72a3d02c4974d8e8"} Mar 17 09:24:15 crc kubenswrapper[4813]: I0317 09:24:15.098957 4813 scope.go:117] "RemoveContainer" containerID="e0ba14bb710e8ef535a87d43e489a6972a73ab00b1733aa05cbdc8da2506f48b" Mar 17 09:24:15 crc kubenswrapper[4813]: W0317 09:24:15.180517 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9a6501c3_1a60_4c0d_b190_5a8c4335b527.slice/crio-11cad663d02b4fc138dd6ca203be3b13edff8ec07d719e19d497b29352913fb5 WatchSource:0}: Error finding container 11cad663d02b4fc138dd6ca203be3b13edff8ec07d719e19d497b29352913fb5: Status 404 returned error can't find the container with id 11cad663d02b4fc138dd6ca203be3b13edff8ec07d719e19d497b29352913fb5 Mar 17 09:24:15 crc kubenswrapper[4813]: I0317 09:24:15.181431 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562324-zfp5g"] Mar 17 09:24:16 crc kubenswrapper[4813]: I0317 09:24:16.109040 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562324-zfp5g" event={"ID":"9a6501c3-1a60-4c0d-b190-5a8c4335b527","Type":"ContainerStarted","Data":"11cad663d02b4fc138dd6ca203be3b13edff8ec07d719e19d497b29352913fb5"} Mar 17 09:24:17 crc kubenswrapper[4813]: I0317 09:24:17.117629 4813 generic.go:334] "Generic (PLEG): container finished" podID="9a6501c3-1a60-4c0d-b190-5a8c4335b527" containerID="e7d340a3db815657bdf95e7b75ba6d139f9ce745fe0d402a1d6530f938945399" exitCode=0 Mar 17 09:24:17 crc kubenswrapper[4813]: I0317 09:24:17.117743 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562324-zfp5g" event={"ID":"9a6501c3-1a60-4c0d-b190-5a8c4335b527","Type":"ContainerDied","Data":"e7d340a3db815657bdf95e7b75ba6d139f9ce745fe0d402a1d6530f938945399"} Mar 17 09:24:18 crc kubenswrapper[4813]: I0317 09:24:18.487141 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562324-zfp5g" Mar 17 09:24:18 crc kubenswrapper[4813]: I0317 09:24:18.519699 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzx8g\" (UniqueName: \"kubernetes.io/projected/9a6501c3-1a60-4c0d-b190-5a8c4335b527-kube-api-access-fzx8g\") pod \"9a6501c3-1a60-4c0d-b190-5a8c4335b527\" (UID: \"9a6501c3-1a60-4c0d-b190-5a8c4335b527\") " Mar 17 09:24:18 crc kubenswrapper[4813]: I0317 09:24:18.528200 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a6501c3-1a60-4c0d-b190-5a8c4335b527-kube-api-access-fzx8g" (OuterVolumeSpecName: "kube-api-access-fzx8g") pod "9a6501c3-1a60-4c0d-b190-5a8c4335b527" (UID: "9a6501c3-1a60-4c0d-b190-5a8c4335b527"). InnerVolumeSpecName "kube-api-access-fzx8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:24:18 crc kubenswrapper[4813]: I0317 09:24:18.620983 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzx8g\" (UniqueName: \"kubernetes.io/projected/9a6501c3-1a60-4c0d-b190-5a8c4335b527-kube-api-access-fzx8g\") on node \"crc\" DevicePath \"\"" Mar 17 09:24:19 crc kubenswrapper[4813]: I0317 09:24:19.138250 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562324-zfp5g" event={"ID":"9a6501c3-1a60-4c0d-b190-5a8c4335b527","Type":"ContainerDied","Data":"11cad663d02b4fc138dd6ca203be3b13edff8ec07d719e19d497b29352913fb5"} Mar 17 09:24:19 crc kubenswrapper[4813]: I0317 09:24:19.138737 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11cad663d02b4fc138dd6ca203be3b13edff8ec07d719e19d497b29352913fb5" Mar 17 09:24:19 crc kubenswrapper[4813]: I0317 09:24:19.138318 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562324-zfp5g" Mar 17 09:24:19 crc kubenswrapper[4813]: I0317 09:24:19.575203 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562318-l68gh"] Mar 17 09:24:19 crc kubenswrapper[4813]: I0317 09:24:19.582769 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562318-l68gh"] Mar 17 09:24:20 crc kubenswrapper[4813]: I0317 09:24:20.742084 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b2354e2-d41f-47f9-9fd9-11f51b893a3a" path="/var/lib/kubelet/pods/5b2354e2-d41f-47f9-9fd9-11f51b893a3a/volumes" Mar 17 09:24:26 crc kubenswrapper[4813]: I0317 09:24:26.031941 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-slmhn"] Mar 17 09:24:26 crc kubenswrapper[4813]: E0317 09:24:26.032943 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a6501c3-1a60-4c0d-b190-5a8c4335b527" containerName="oc" Mar 17 09:24:26 crc kubenswrapper[4813]: I0317 09:24:26.032958 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a6501c3-1a60-4c0d-b190-5a8c4335b527" containerName="oc" Mar 17 09:24:26 crc kubenswrapper[4813]: I0317 09:24:26.033070 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a6501c3-1a60-4c0d-b190-5a8c4335b527" containerName="oc" Mar 17 09:24:26 crc kubenswrapper[4813]: I0317 09:24:26.033882 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-slmhn" Mar 17 09:24:26 crc kubenswrapper[4813]: I0317 09:24:26.055349 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-slmhn"] Mar 17 09:24:26 crc kubenswrapper[4813]: I0317 09:24:26.225909 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/048cf9db-05e6-4e8c-bbd5-87abfc091726-utilities\") pod \"community-operators-slmhn\" (UID: \"048cf9db-05e6-4e8c-bbd5-87abfc091726\") " pod="openshift-marketplace/community-operators-slmhn" Mar 17 09:24:26 crc kubenswrapper[4813]: I0317 09:24:26.226165 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z4v9z\" (UniqueName: \"kubernetes.io/projected/048cf9db-05e6-4e8c-bbd5-87abfc091726-kube-api-access-z4v9z\") pod \"community-operators-slmhn\" (UID: \"048cf9db-05e6-4e8c-bbd5-87abfc091726\") " pod="openshift-marketplace/community-operators-slmhn" Mar 17 09:24:26 crc kubenswrapper[4813]: I0317 09:24:26.226301 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/048cf9db-05e6-4e8c-bbd5-87abfc091726-catalog-content\") pod \"community-operators-slmhn\" (UID: \"048cf9db-05e6-4e8c-bbd5-87abfc091726\") " pod="openshift-marketplace/community-operators-slmhn" Mar 17 09:24:26 crc kubenswrapper[4813]: I0317 09:24:26.308194 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-jvl96" Mar 17 09:24:26 crc kubenswrapper[4813]: I0317 09:24:26.327171 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/048cf9db-05e6-4e8c-bbd5-87abfc091726-utilities\") pod \"community-operators-slmhn\" (UID: \"048cf9db-05e6-4e8c-bbd5-87abfc091726\") " pod="openshift-marketplace/community-operators-slmhn" Mar 17 09:24:26 crc kubenswrapper[4813]: I0317 09:24:26.327252 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z4v9z\" (UniqueName: \"kubernetes.io/projected/048cf9db-05e6-4e8c-bbd5-87abfc091726-kube-api-access-z4v9z\") pod \"community-operators-slmhn\" (UID: \"048cf9db-05e6-4e8c-bbd5-87abfc091726\") " pod="openshift-marketplace/community-operators-slmhn" Mar 17 09:24:26 crc kubenswrapper[4813]: I0317 09:24:26.327296 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/048cf9db-05e6-4e8c-bbd5-87abfc091726-catalog-content\") pod \"community-operators-slmhn\" (UID: \"048cf9db-05e6-4e8c-bbd5-87abfc091726\") " pod="openshift-marketplace/community-operators-slmhn" Mar 17 09:24:26 crc kubenswrapper[4813]: I0317 09:24:26.327686 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/048cf9db-05e6-4e8c-bbd5-87abfc091726-utilities\") pod \"community-operators-slmhn\" (UID: \"048cf9db-05e6-4e8c-bbd5-87abfc091726\") " pod="openshift-marketplace/community-operators-slmhn" Mar 17 09:24:26 crc kubenswrapper[4813]: I0317 09:24:26.328131 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/048cf9db-05e6-4e8c-bbd5-87abfc091726-catalog-content\") pod \"community-operators-slmhn\" (UID: \"048cf9db-05e6-4e8c-bbd5-87abfc091726\") " pod="openshift-marketplace/community-operators-slmhn" Mar 17 09:24:26 crc kubenswrapper[4813]: I0317 09:24:26.360909 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z4v9z\" (UniqueName: \"kubernetes.io/projected/048cf9db-05e6-4e8c-bbd5-87abfc091726-kube-api-access-z4v9z\") pod \"community-operators-slmhn\" (UID: \"048cf9db-05e6-4e8c-bbd5-87abfc091726\") " pod="openshift-marketplace/community-operators-slmhn" Mar 17 09:24:26 crc kubenswrapper[4813]: I0317 09:24:26.363146 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-slmhn" Mar 17 09:24:26 crc kubenswrapper[4813]: I0317 09:24:26.627441 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-slmhn"] Mar 17 09:24:27 crc kubenswrapper[4813]: I0317 09:24:27.198889 4813 generic.go:334] "Generic (PLEG): container finished" podID="048cf9db-05e6-4e8c-bbd5-87abfc091726" containerID="e12bf458d0f830396e2300ce6aac552a6d2ddffd124929fd5cdd0e39786a626d" exitCode=0 Mar 17 09:24:27 crc kubenswrapper[4813]: I0317 09:24:27.198988 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-slmhn" event={"ID":"048cf9db-05e6-4e8c-bbd5-87abfc091726","Type":"ContainerDied","Data":"e12bf458d0f830396e2300ce6aac552a6d2ddffd124929fd5cdd0e39786a626d"} Mar 17 09:24:27 crc kubenswrapper[4813]: I0317 09:24:27.199271 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-slmhn" event={"ID":"048cf9db-05e6-4e8c-bbd5-87abfc091726","Type":"ContainerStarted","Data":"afb62a9bd785c8370559d0c4ccc3a63f279a749eaa7322cbaa2e0c6da016d0e6"} Mar 17 09:24:28 crc kubenswrapper[4813]: I0317 09:24:28.208450 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-slmhn" event={"ID":"048cf9db-05e6-4e8c-bbd5-87abfc091726","Type":"ContainerStarted","Data":"90453e0ae021ec25703d39f45f6c0cac0d0dabbf47ab588ab98fa98ded94bc37"} Mar 17 09:24:29 crc kubenswrapper[4813]: I0317 09:24:29.219626 4813 generic.go:334] "Generic (PLEG): container finished" podID="048cf9db-05e6-4e8c-bbd5-87abfc091726" containerID="90453e0ae021ec25703d39f45f6c0cac0d0dabbf47ab588ab98fa98ded94bc37" exitCode=0 Mar 17 09:24:29 crc kubenswrapper[4813]: I0317 09:24:29.219664 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-slmhn" event={"ID":"048cf9db-05e6-4e8c-bbd5-87abfc091726","Type":"ContainerDied","Data":"90453e0ae021ec25703d39f45f6c0cac0d0dabbf47ab588ab98fa98ded94bc37"} Mar 17 09:24:30 crc kubenswrapper[4813]: I0317 09:24:30.231257 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-slmhn" event={"ID":"048cf9db-05e6-4e8c-bbd5-87abfc091726","Type":"ContainerStarted","Data":"735aa62d3d9d52f3385e4cb5145e9946aed41cae77b12ad5145198d9413beec9"} Mar 17 09:24:30 crc kubenswrapper[4813]: I0317 09:24:30.262434 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-slmhn" podStartSLOduration=1.767790678 podStartE2EDuration="4.262401323s" podCreationTimestamp="2026-03-17 09:24:26 +0000 UTC" firstStartedPulling="2026-03-17 09:24:27.200895812 +0000 UTC m=+889.301699341" lastFinishedPulling="2026-03-17 09:24:29.695506477 +0000 UTC m=+891.796309986" observedRunningTime="2026-03-17 09:24:30.253399741 +0000 UTC m=+892.354203280" watchObservedRunningTime="2026-03-17 09:24:30.262401323 +0000 UTC m=+892.363204892" Mar 17 09:24:33 crc kubenswrapper[4813]: I0317 09:24:33.061890 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf"] Mar 17 09:24:33 crc kubenswrapper[4813]: I0317 09:24:33.064861 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf" Mar 17 09:24:33 crc kubenswrapper[4813]: I0317 09:24:33.068262 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 17 09:24:33 crc kubenswrapper[4813]: I0317 09:24:33.079265 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf"] Mar 17 09:24:33 crc kubenswrapper[4813]: I0317 09:24:33.121581 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb556790-2a96-412b-9847-3cd62873bfed-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf\" (UID: \"bb556790-2a96-412b-9847-3cd62873bfed\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf" Mar 17 09:24:33 crc kubenswrapper[4813]: I0317 09:24:33.121675 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lckb6\" (UniqueName: \"kubernetes.io/projected/bb556790-2a96-412b-9847-3cd62873bfed-kube-api-access-lckb6\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf\" (UID: \"bb556790-2a96-412b-9847-3cd62873bfed\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf" Mar 17 09:24:33 crc kubenswrapper[4813]: I0317 09:24:33.121793 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb556790-2a96-412b-9847-3cd62873bfed-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf\" (UID: \"bb556790-2a96-412b-9847-3cd62873bfed\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf" Mar 17 09:24:33 crc kubenswrapper[4813]: I0317 09:24:33.222767 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb556790-2a96-412b-9847-3cd62873bfed-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf\" (UID: \"bb556790-2a96-412b-9847-3cd62873bfed\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf" Mar 17 09:24:33 crc kubenswrapper[4813]: I0317 09:24:33.223103 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb556790-2a96-412b-9847-3cd62873bfed-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf\" (UID: \"bb556790-2a96-412b-9847-3cd62873bfed\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf" Mar 17 09:24:33 crc kubenswrapper[4813]: I0317 09:24:33.223132 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lckb6\" (UniqueName: \"kubernetes.io/projected/bb556790-2a96-412b-9847-3cd62873bfed-kube-api-access-lckb6\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf\" (UID: \"bb556790-2a96-412b-9847-3cd62873bfed\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf" Mar 17 09:24:33 crc kubenswrapper[4813]: I0317 09:24:33.223728 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb556790-2a96-412b-9847-3cd62873bfed-util\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf\" (UID: \"bb556790-2a96-412b-9847-3cd62873bfed\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf" Mar 17 09:24:33 crc kubenswrapper[4813]: I0317 09:24:33.223835 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb556790-2a96-412b-9847-3cd62873bfed-bundle\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf\" (UID: \"bb556790-2a96-412b-9847-3cd62873bfed\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf" Mar 17 09:24:33 crc kubenswrapper[4813]: I0317 09:24:33.263373 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lckb6\" (UniqueName: \"kubernetes.io/projected/bb556790-2a96-412b-9847-3cd62873bfed-kube-api-access-lckb6\") pod \"1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf\" (UID: \"bb556790-2a96-412b-9847-3cd62873bfed\") " pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf" Mar 17 09:24:33 crc kubenswrapper[4813]: I0317 09:24:33.397153 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf" Mar 17 09:24:33 crc kubenswrapper[4813]: I0317 09:24:33.712439 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf"] Mar 17 09:24:33 crc kubenswrapper[4813]: W0317 09:24:33.722041 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbb556790_2a96_412b_9847_3cd62873bfed.slice/crio-09428237097d5fe0ef0a93ba876d97c065159b1ccd6eda6c1c77e06f76c5e1c1 WatchSource:0}: Error finding container 09428237097d5fe0ef0a93ba876d97c065159b1ccd6eda6c1c77e06f76c5e1c1: Status 404 returned error can't find the container with id 09428237097d5fe0ef0a93ba876d97c065159b1ccd6eda6c1c77e06f76c5e1c1 Mar 17 09:24:34 crc kubenswrapper[4813]: I0317 09:24:34.260772 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf" event={"ID":"bb556790-2a96-412b-9847-3cd62873bfed","Type":"ContainerStarted","Data":"b4725d28f9969d3e5a5190a7c1d6717a4357db7a7c303580af6da32cc743f8da"} Mar 17 09:24:34 crc kubenswrapper[4813]: I0317 09:24:34.260847 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf" event={"ID":"bb556790-2a96-412b-9847-3cd62873bfed","Type":"ContainerStarted","Data":"09428237097d5fe0ef0a93ba876d97c065159b1ccd6eda6c1c77e06f76c5e1c1"} Mar 17 09:24:35 crc kubenswrapper[4813]: I0317 09:24:35.272421 4813 generic.go:334] "Generic (PLEG): container finished" podID="bb556790-2a96-412b-9847-3cd62873bfed" containerID="b4725d28f9969d3e5a5190a7c1d6717a4357db7a7c303580af6da32cc743f8da" exitCode=0 Mar 17 09:24:35 crc kubenswrapper[4813]: I0317 09:24:35.272527 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf" event={"ID":"bb556790-2a96-412b-9847-3cd62873bfed","Type":"ContainerDied","Data":"b4725d28f9969d3e5a5190a7c1d6717a4357db7a7c303580af6da32cc743f8da"} Mar 17 09:24:35 crc kubenswrapper[4813]: I0317 09:24:35.613445 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vf8rq"] Mar 17 09:24:35 crc kubenswrapper[4813]: I0317 09:24:35.614545 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vf8rq" Mar 17 09:24:35 crc kubenswrapper[4813]: I0317 09:24:35.618367 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vf8rq"] Mar 17 09:24:35 crc kubenswrapper[4813]: I0317 09:24:35.661795 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fc7a47b-182b-40dc-bf52-752b7434f448-catalog-content\") pod \"redhat-operators-vf8rq\" (UID: \"4fc7a47b-182b-40dc-bf52-752b7434f448\") " pod="openshift-marketplace/redhat-operators-vf8rq" Mar 17 09:24:35 crc kubenswrapper[4813]: I0317 09:24:35.661939 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f64h4\" (UniqueName: \"kubernetes.io/projected/4fc7a47b-182b-40dc-bf52-752b7434f448-kube-api-access-f64h4\") pod \"redhat-operators-vf8rq\" (UID: \"4fc7a47b-182b-40dc-bf52-752b7434f448\") " pod="openshift-marketplace/redhat-operators-vf8rq" Mar 17 09:24:35 crc kubenswrapper[4813]: I0317 09:24:35.662011 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fc7a47b-182b-40dc-bf52-752b7434f448-utilities\") pod \"redhat-operators-vf8rq\" (UID: \"4fc7a47b-182b-40dc-bf52-752b7434f448\") " pod="openshift-marketplace/redhat-operators-vf8rq" Mar 17 09:24:35 crc kubenswrapper[4813]: I0317 09:24:35.763514 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fc7a47b-182b-40dc-bf52-752b7434f448-catalog-content\") pod \"redhat-operators-vf8rq\" (UID: \"4fc7a47b-182b-40dc-bf52-752b7434f448\") " pod="openshift-marketplace/redhat-operators-vf8rq" Mar 17 09:24:35 crc kubenswrapper[4813]: I0317 09:24:35.763626 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f64h4\" (UniqueName: \"kubernetes.io/projected/4fc7a47b-182b-40dc-bf52-752b7434f448-kube-api-access-f64h4\") pod \"redhat-operators-vf8rq\" (UID: \"4fc7a47b-182b-40dc-bf52-752b7434f448\") " pod="openshift-marketplace/redhat-operators-vf8rq" Mar 17 09:24:35 crc kubenswrapper[4813]: I0317 09:24:35.763663 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fc7a47b-182b-40dc-bf52-752b7434f448-utilities\") pod \"redhat-operators-vf8rq\" (UID: \"4fc7a47b-182b-40dc-bf52-752b7434f448\") " pod="openshift-marketplace/redhat-operators-vf8rq" Mar 17 09:24:35 crc kubenswrapper[4813]: I0317 09:24:35.764629 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4fc7a47b-182b-40dc-bf52-752b7434f448-utilities\") pod \"redhat-operators-vf8rq\" (UID: \"4fc7a47b-182b-40dc-bf52-752b7434f448\") " pod="openshift-marketplace/redhat-operators-vf8rq" Mar 17 09:24:35 crc kubenswrapper[4813]: I0317 09:24:35.764762 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4fc7a47b-182b-40dc-bf52-752b7434f448-catalog-content\") pod \"redhat-operators-vf8rq\" (UID: \"4fc7a47b-182b-40dc-bf52-752b7434f448\") " pod="openshift-marketplace/redhat-operators-vf8rq" Mar 17 09:24:35 crc kubenswrapper[4813]: I0317 09:24:35.792012 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f64h4\" (UniqueName: \"kubernetes.io/projected/4fc7a47b-182b-40dc-bf52-752b7434f448-kube-api-access-f64h4\") pod \"redhat-operators-vf8rq\" (UID: \"4fc7a47b-182b-40dc-bf52-752b7434f448\") " pod="openshift-marketplace/redhat-operators-vf8rq" Mar 17 09:24:35 crc kubenswrapper[4813]: I0317 09:24:35.945772 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vf8rq" Mar 17 09:24:36 crc kubenswrapper[4813]: I0317 09:24:36.143416 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vf8rq"] Mar 17 09:24:36 crc kubenswrapper[4813]: I0317 09:24:36.279959 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vf8rq" event={"ID":"4fc7a47b-182b-40dc-bf52-752b7434f448","Type":"ContainerStarted","Data":"c03473ae6a868d36ae254adbf10ca74e7dc844f97c8bf4c741da66919df0c316"} Mar 17 09:24:36 crc kubenswrapper[4813]: I0317 09:24:36.365091 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-slmhn" Mar 17 09:24:36 crc kubenswrapper[4813]: I0317 09:24:36.365140 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-slmhn" Mar 17 09:24:36 crc kubenswrapper[4813]: I0317 09:24:36.402237 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-slmhn" Mar 17 09:24:37 crc kubenswrapper[4813]: I0317 09:24:37.286856 4813 generic.go:334] "Generic (PLEG): container finished" podID="4fc7a47b-182b-40dc-bf52-752b7434f448" containerID="861e7ea5e61cb856856dbd0ec0193db94fd26d8752ca6393c8bd6f09c22ac922" exitCode=0 Mar 17 09:24:37 crc kubenswrapper[4813]: I0317 09:24:37.286912 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vf8rq" event={"ID":"4fc7a47b-182b-40dc-bf52-752b7434f448","Type":"ContainerDied","Data":"861e7ea5e61cb856856dbd0ec0193db94fd26d8752ca6393c8bd6f09c22ac922"} Mar 17 09:24:37 crc kubenswrapper[4813]: I0317 09:24:37.289063 4813 generic.go:334] "Generic (PLEG): container finished" podID="bb556790-2a96-412b-9847-3cd62873bfed" containerID="0fc046d2c6edb14310e92b737ce0fad640e8738304a5c23535e7ae468642787e" exitCode=0 Mar 17 09:24:37 crc kubenswrapper[4813]: I0317 09:24:37.289746 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf" event={"ID":"bb556790-2a96-412b-9847-3cd62873bfed","Type":"ContainerDied","Data":"0fc046d2c6edb14310e92b737ce0fad640e8738304a5c23535e7ae468642787e"} Mar 17 09:24:37 crc kubenswrapper[4813]: I0317 09:24:37.343906 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-slmhn" Mar 17 09:24:38 crc kubenswrapper[4813]: I0317 09:24:38.300797 4813 generic.go:334] "Generic (PLEG): container finished" podID="bb556790-2a96-412b-9847-3cd62873bfed" containerID="5b7ada8c5ce89da16f009628794af6d29834e2ad20f2608e968b576d997f08a7" exitCode=0 Mar 17 09:24:38 crc kubenswrapper[4813]: I0317 09:24:38.300877 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf" event={"ID":"bb556790-2a96-412b-9847-3cd62873bfed","Type":"ContainerDied","Data":"5b7ada8c5ce89da16f009628794af6d29834e2ad20f2608e968b576d997f08a7"} Mar 17 09:24:39 crc kubenswrapper[4813]: I0317 09:24:39.297562 4813 scope.go:117] "RemoveContainer" containerID="dad746fa90ef589dad6ffc30fda9d259a6ebb043370f2f9ce8eff8745df59eb2" Mar 17 09:24:39 crc kubenswrapper[4813]: I0317 09:24:39.549310 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf" Mar 17 09:24:39 crc kubenswrapper[4813]: I0317 09:24:39.740850 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb556790-2a96-412b-9847-3cd62873bfed-bundle\") pod \"bb556790-2a96-412b-9847-3cd62873bfed\" (UID: \"bb556790-2a96-412b-9847-3cd62873bfed\") " Mar 17 09:24:39 crc kubenswrapper[4813]: I0317 09:24:39.740952 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lckb6\" (UniqueName: \"kubernetes.io/projected/bb556790-2a96-412b-9847-3cd62873bfed-kube-api-access-lckb6\") pod \"bb556790-2a96-412b-9847-3cd62873bfed\" (UID: \"bb556790-2a96-412b-9847-3cd62873bfed\") " Mar 17 09:24:39 crc kubenswrapper[4813]: I0317 09:24:39.740973 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb556790-2a96-412b-9847-3cd62873bfed-util\") pod \"bb556790-2a96-412b-9847-3cd62873bfed\" (UID: \"bb556790-2a96-412b-9847-3cd62873bfed\") " Mar 17 09:24:39 crc kubenswrapper[4813]: I0317 09:24:39.741394 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb556790-2a96-412b-9847-3cd62873bfed-bundle" (OuterVolumeSpecName: "bundle") pod "bb556790-2a96-412b-9847-3cd62873bfed" (UID: "bb556790-2a96-412b-9847-3cd62873bfed"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:24:39 crc kubenswrapper[4813]: I0317 09:24:39.746766 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb556790-2a96-412b-9847-3cd62873bfed-kube-api-access-lckb6" (OuterVolumeSpecName: "kube-api-access-lckb6") pod "bb556790-2a96-412b-9847-3cd62873bfed" (UID: "bb556790-2a96-412b-9847-3cd62873bfed"). InnerVolumeSpecName "kube-api-access-lckb6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:24:39 crc kubenswrapper[4813]: I0317 09:24:39.750474 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb556790-2a96-412b-9847-3cd62873bfed-util" (OuterVolumeSpecName: "util") pod "bb556790-2a96-412b-9847-3cd62873bfed" (UID: "bb556790-2a96-412b-9847-3cd62873bfed"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:24:39 crc kubenswrapper[4813]: I0317 09:24:39.842214 4813 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bb556790-2a96-412b-9847-3cd62873bfed-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:24:39 crc kubenswrapper[4813]: I0317 09:24:39.842827 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lckb6\" (UniqueName: \"kubernetes.io/projected/bb556790-2a96-412b-9847-3cd62873bfed-kube-api-access-lckb6\") on node \"crc\" DevicePath \"\"" Mar 17 09:24:39 crc kubenswrapper[4813]: I0317 09:24:39.842845 4813 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bb556790-2a96-412b-9847-3cd62873bfed-util\") on node \"crc\" DevicePath \"\"" Mar 17 09:24:39 crc kubenswrapper[4813]: I0317 09:24:39.990072 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-slmhn"] Mar 17 09:24:39 crc kubenswrapper[4813]: I0317 09:24:39.990300 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-slmhn" podUID="048cf9db-05e6-4e8c-bbd5-87abfc091726" containerName="registry-server" containerID="cri-o://735aa62d3d9d52f3385e4cb5145e9946aed41cae77b12ad5145198d9413beec9" gracePeriod=2 Mar 17 09:24:40 crc kubenswrapper[4813]: I0317 09:24:40.317297 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf" event={"ID":"bb556790-2a96-412b-9847-3cd62873bfed","Type":"ContainerDied","Data":"09428237097d5fe0ef0a93ba876d97c065159b1ccd6eda6c1c77e06f76c5e1c1"} Mar 17 09:24:40 crc kubenswrapper[4813]: I0317 09:24:40.317349 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="09428237097d5fe0ef0a93ba876d97c065159b1ccd6eda6c1c77e06f76c5e1c1" Mar 17 09:24:40 crc kubenswrapper[4813]: I0317 09:24:40.317430 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf" Mar 17 09:24:40 crc kubenswrapper[4813]: I0317 09:24:40.324939 4813 generic.go:334] "Generic (PLEG): container finished" podID="048cf9db-05e6-4e8c-bbd5-87abfc091726" containerID="735aa62d3d9d52f3385e4cb5145e9946aed41cae77b12ad5145198d9413beec9" exitCode=0 Mar 17 09:24:40 crc kubenswrapper[4813]: I0317 09:24:40.324991 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-slmhn" event={"ID":"048cf9db-05e6-4e8c-bbd5-87abfc091726","Type":"ContainerDied","Data":"735aa62d3d9d52f3385e4cb5145e9946aed41cae77b12ad5145198d9413beec9"} Mar 17 09:24:40 crc kubenswrapper[4813]: I0317 09:24:40.472254 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-slmhn" Mar 17 09:24:40 crc kubenswrapper[4813]: I0317 09:24:40.652436 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z4v9z\" (UniqueName: \"kubernetes.io/projected/048cf9db-05e6-4e8c-bbd5-87abfc091726-kube-api-access-z4v9z\") pod \"048cf9db-05e6-4e8c-bbd5-87abfc091726\" (UID: \"048cf9db-05e6-4e8c-bbd5-87abfc091726\") " Mar 17 09:24:40 crc kubenswrapper[4813]: I0317 09:24:40.653457 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/048cf9db-05e6-4e8c-bbd5-87abfc091726-catalog-content\") pod \"048cf9db-05e6-4e8c-bbd5-87abfc091726\" (UID: \"048cf9db-05e6-4e8c-bbd5-87abfc091726\") " Mar 17 09:24:40 crc kubenswrapper[4813]: I0317 09:24:40.657879 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/048cf9db-05e6-4e8c-bbd5-87abfc091726-utilities\") pod \"048cf9db-05e6-4e8c-bbd5-87abfc091726\" (UID: \"048cf9db-05e6-4e8c-bbd5-87abfc091726\") " Mar 17 09:24:40 crc kubenswrapper[4813]: I0317 09:24:40.658091 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/048cf9db-05e6-4e8c-bbd5-87abfc091726-kube-api-access-z4v9z" (OuterVolumeSpecName: "kube-api-access-z4v9z") pod "048cf9db-05e6-4e8c-bbd5-87abfc091726" (UID: "048cf9db-05e6-4e8c-bbd5-87abfc091726"). InnerVolumeSpecName "kube-api-access-z4v9z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:24:40 crc kubenswrapper[4813]: I0317 09:24:40.658640 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z4v9z\" (UniqueName: \"kubernetes.io/projected/048cf9db-05e6-4e8c-bbd5-87abfc091726-kube-api-access-z4v9z\") on node \"crc\" DevicePath \"\"" Mar 17 09:24:40 crc kubenswrapper[4813]: I0317 09:24:40.658687 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/048cf9db-05e6-4e8c-bbd5-87abfc091726-utilities" (OuterVolumeSpecName: "utilities") pod "048cf9db-05e6-4e8c-bbd5-87abfc091726" (UID: "048cf9db-05e6-4e8c-bbd5-87abfc091726"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:24:40 crc kubenswrapper[4813]: I0317 09:24:40.699851 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/048cf9db-05e6-4e8c-bbd5-87abfc091726-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "048cf9db-05e6-4e8c-bbd5-87abfc091726" (UID: "048cf9db-05e6-4e8c-bbd5-87abfc091726"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:24:40 crc kubenswrapper[4813]: I0317 09:24:40.759787 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/048cf9db-05e6-4e8c-bbd5-87abfc091726-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:24:40 crc kubenswrapper[4813]: I0317 09:24:40.759814 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/048cf9db-05e6-4e8c-bbd5-87abfc091726-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:24:41 crc kubenswrapper[4813]: I0317 09:24:41.334303 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-slmhn" event={"ID":"048cf9db-05e6-4e8c-bbd5-87abfc091726","Type":"ContainerDied","Data":"afb62a9bd785c8370559d0c4ccc3a63f279a749eaa7322cbaa2e0c6da016d0e6"} Mar 17 09:24:41 crc kubenswrapper[4813]: I0317 09:24:41.334695 4813 scope.go:117] "RemoveContainer" containerID="735aa62d3d9d52f3385e4cb5145e9946aed41cae77b12ad5145198d9413beec9" Mar 17 09:24:41 crc kubenswrapper[4813]: I0317 09:24:41.334432 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-slmhn" Mar 17 09:24:41 crc kubenswrapper[4813]: I0317 09:24:41.353035 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-slmhn"] Mar 17 09:24:41 crc kubenswrapper[4813]: I0317 09:24:41.359187 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-slmhn"] Mar 17 09:24:42 crc kubenswrapper[4813]: I0317 09:24:42.751067 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="048cf9db-05e6-4e8c-bbd5-87abfc091726" path="/var/lib/kubelet/pods/048cf9db-05e6-4e8c-bbd5-87abfc091726/volumes" Mar 17 09:24:43 crc kubenswrapper[4813]: I0317 09:24:43.684972 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-77p2p"] Mar 17 09:24:43 crc kubenswrapper[4813]: E0317 09:24:43.685848 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="048cf9db-05e6-4e8c-bbd5-87abfc091726" containerName="registry-server" Mar 17 09:24:43 crc kubenswrapper[4813]: I0317 09:24:43.685867 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="048cf9db-05e6-4e8c-bbd5-87abfc091726" containerName="registry-server" Mar 17 09:24:43 crc kubenswrapper[4813]: E0317 09:24:43.685890 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb556790-2a96-412b-9847-3cd62873bfed" containerName="pull" Mar 17 09:24:43 crc kubenswrapper[4813]: I0317 09:24:43.685899 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb556790-2a96-412b-9847-3cd62873bfed" containerName="pull" Mar 17 09:24:43 crc kubenswrapper[4813]: E0317 09:24:43.685917 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="048cf9db-05e6-4e8c-bbd5-87abfc091726" containerName="extract-utilities" Mar 17 09:24:43 crc kubenswrapper[4813]: I0317 09:24:43.685925 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="048cf9db-05e6-4e8c-bbd5-87abfc091726" containerName="extract-utilities" Mar 17 09:24:43 crc kubenswrapper[4813]: E0317 09:24:43.685936 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="048cf9db-05e6-4e8c-bbd5-87abfc091726" containerName="extract-content" Mar 17 09:24:43 crc kubenswrapper[4813]: I0317 09:24:43.685944 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="048cf9db-05e6-4e8c-bbd5-87abfc091726" containerName="extract-content" Mar 17 09:24:43 crc kubenswrapper[4813]: E0317 09:24:43.685954 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb556790-2a96-412b-9847-3cd62873bfed" containerName="util" Mar 17 09:24:43 crc kubenswrapper[4813]: I0317 09:24:43.685961 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb556790-2a96-412b-9847-3cd62873bfed" containerName="util" Mar 17 09:24:43 crc kubenswrapper[4813]: E0317 09:24:43.685969 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb556790-2a96-412b-9847-3cd62873bfed" containerName="extract" Mar 17 09:24:43 crc kubenswrapper[4813]: I0317 09:24:43.685976 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb556790-2a96-412b-9847-3cd62873bfed" containerName="extract" Mar 17 09:24:43 crc kubenswrapper[4813]: I0317 09:24:43.686093 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb556790-2a96-412b-9847-3cd62873bfed" containerName="extract" Mar 17 09:24:43 crc kubenswrapper[4813]: I0317 09:24:43.686106 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="048cf9db-05e6-4e8c-bbd5-87abfc091726" containerName="registry-server" Mar 17 09:24:43 crc kubenswrapper[4813]: I0317 09:24:43.686568 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-77p2p" Mar 17 09:24:43 crc kubenswrapper[4813]: I0317 09:24:43.688817 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-bw5n5" Mar 17 09:24:43 crc kubenswrapper[4813]: I0317 09:24:43.689234 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Mar 17 09:24:43 crc kubenswrapper[4813]: I0317 09:24:43.689479 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Mar 17 09:24:43 crc kubenswrapper[4813]: I0317 09:24:43.694820 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-77p2p"] Mar 17 09:24:43 crc kubenswrapper[4813]: I0317 09:24:43.717919 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr489\" (UniqueName: \"kubernetes.io/projected/acc2af0d-b40c-48d9-9b83-d0cf7755c37b-kube-api-access-tr489\") pod \"nmstate-operator-796d4cfff4-77p2p\" (UID: \"acc2af0d-b40c-48d9-9b83-d0cf7755c37b\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-77p2p" Mar 17 09:24:43 crc kubenswrapper[4813]: I0317 09:24:43.818862 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr489\" (UniqueName: \"kubernetes.io/projected/acc2af0d-b40c-48d9-9b83-d0cf7755c37b-kube-api-access-tr489\") pod \"nmstate-operator-796d4cfff4-77p2p\" (UID: \"acc2af0d-b40c-48d9-9b83-d0cf7755c37b\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-77p2p" Mar 17 09:24:43 crc kubenswrapper[4813]: I0317 09:24:43.836226 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr489\" (UniqueName: \"kubernetes.io/projected/acc2af0d-b40c-48d9-9b83-d0cf7755c37b-kube-api-access-tr489\") pod \"nmstate-operator-796d4cfff4-77p2p\" (UID: \"acc2af0d-b40c-48d9-9b83-d0cf7755c37b\") " pod="openshift-nmstate/nmstate-operator-796d4cfff4-77p2p" Mar 17 09:24:44 crc kubenswrapper[4813]: I0317 09:24:44.009445 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-796d4cfff4-77p2p" Mar 17 09:24:44 crc kubenswrapper[4813]: I0317 09:24:44.904860 4813 scope.go:117] "RemoveContainer" containerID="90453e0ae021ec25703d39f45f6c0cac0d0dabbf47ab588ab98fa98ded94bc37" Mar 17 09:24:45 crc kubenswrapper[4813]: I0317 09:24:45.034540 4813 scope.go:117] "RemoveContainer" containerID="e12bf458d0f830396e2300ce6aac552a6d2ddffd124929fd5cdd0e39786a626d" Mar 17 09:24:45 crc kubenswrapper[4813]: I0317 09:24:45.347433 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-796d4cfff4-77p2p"] Mar 17 09:24:45 crc kubenswrapper[4813]: I0317 09:24:45.361102 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vf8rq" event={"ID":"4fc7a47b-182b-40dc-bf52-752b7434f448","Type":"ContainerStarted","Data":"21fdc177d8f24afa60180b95f2aede2fdb1635d10c92937037b8ce71b84595dd"} Mar 17 09:24:46 crc kubenswrapper[4813]: I0317 09:24:46.369635 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-77p2p" event={"ID":"acc2af0d-b40c-48d9-9b83-d0cf7755c37b","Type":"ContainerStarted","Data":"65815252961afa8ec651f1269b2ef96bd1d0433b100d109ad868dd6c816ef809"} Mar 17 09:24:46 crc kubenswrapper[4813]: I0317 09:24:46.373399 4813 generic.go:334] "Generic (PLEG): container finished" podID="4fc7a47b-182b-40dc-bf52-752b7434f448" containerID="21fdc177d8f24afa60180b95f2aede2fdb1635d10c92937037b8ce71b84595dd" exitCode=0 Mar 17 09:24:46 crc kubenswrapper[4813]: I0317 09:24:46.373471 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vf8rq" event={"ID":"4fc7a47b-182b-40dc-bf52-752b7434f448","Type":"ContainerDied","Data":"21fdc177d8f24afa60180b95f2aede2fdb1635d10c92937037b8ce71b84595dd"} Mar 17 09:24:47 crc kubenswrapper[4813]: I0317 09:24:47.380358 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vf8rq" event={"ID":"4fc7a47b-182b-40dc-bf52-752b7434f448","Type":"ContainerStarted","Data":"fbe2762a1cbbe3b21e66bc7d027bf97aa108671cabe69d0d8c53918c4677fb93"} Mar 17 09:24:47 crc kubenswrapper[4813]: I0317 09:24:47.400957 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vf8rq" podStartSLOduration=2.861129526 podStartE2EDuration="12.400939291s" podCreationTimestamp="2026-03-17 09:24:35 +0000 UTC" firstStartedPulling="2026-03-17 09:24:37.288945454 +0000 UTC m=+899.389748963" lastFinishedPulling="2026-03-17 09:24:46.828755199 +0000 UTC m=+908.929558728" observedRunningTime="2026-03-17 09:24:47.397988777 +0000 UTC m=+909.498792276" watchObservedRunningTime="2026-03-17 09:24:47.400939291 +0000 UTC m=+909.501742790" Mar 17 09:24:48 crc kubenswrapper[4813]: I0317 09:24:48.387805 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-796d4cfff4-77p2p" event={"ID":"acc2af0d-b40c-48d9-9b83-d0cf7755c37b","Type":"ContainerStarted","Data":"e71ac3ae9336e70c750f5b0965d49a004895c13836ce10e3c6fd6b1112dee953"} Mar 17 09:24:48 crc kubenswrapper[4813]: I0317 09:24:48.418043 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-796d4cfff4-77p2p" podStartSLOduration=3.130547106 podStartE2EDuration="5.418018756s" podCreationTimestamp="2026-03-17 09:24:43 +0000 UTC" firstStartedPulling="2026-03-17 09:24:45.357393044 +0000 UTC m=+907.458196543" lastFinishedPulling="2026-03-17 09:24:47.644864694 +0000 UTC m=+909.745668193" observedRunningTime="2026-03-17 09:24:48.410648282 +0000 UTC m=+910.511451821" watchObservedRunningTime="2026-03-17 09:24:48.418018756 +0000 UTC m=+910.518822285" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.645167 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-95n52"] Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.647092 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-95n52" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.649422 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.649725 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-bz4rd" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.650617 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-k2x5n"] Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.651823 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-k2x5n" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.682084 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-zckpn"] Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.682924 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-zckpn" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.689575 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-k2x5n"] Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.718137 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-95n52"] Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.743931 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xbxnm\" (UniqueName: \"kubernetes.io/projected/23256f06-964f-4f6f-bedf-6159a92a06d0-kube-api-access-xbxnm\") pod \"nmstate-webhook-5f558f5558-95n52\" (UID: \"23256f06-964f-4f6f-bedf-6159a92a06d0\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-95n52" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.743975 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/8813052b-cc40-4fe7-b4de-7aeb3832d4bb-dbus-socket\") pod \"nmstate-handler-zckpn\" (UID: \"8813052b-cc40-4fe7-b4de-7aeb3832d4bb\") " pod="openshift-nmstate/nmstate-handler-zckpn" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.744006 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/8813052b-cc40-4fe7-b4de-7aeb3832d4bb-nmstate-lock\") pod \"nmstate-handler-zckpn\" (UID: \"8813052b-cc40-4fe7-b4de-7aeb3832d4bb\") " pod="openshift-nmstate/nmstate-handler-zckpn" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.744022 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/23256f06-964f-4f6f-bedf-6159a92a06d0-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-95n52\" (UID: \"23256f06-964f-4f6f-bedf-6159a92a06d0\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-95n52" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.744143 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/8813052b-cc40-4fe7-b4de-7aeb3832d4bb-ovs-socket\") pod \"nmstate-handler-zckpn\" (UID: \"8813052b-cc40-4fe7-b4de-7aeb3832d4bb\") " pod="openshift-nmstate/nmstate-handler-zckpn" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.744192 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm4k9\" (UniqueName: \"kubernetes.io/projected/0d0f7c9a-ffef-4d0b-ae31-0cfd9e864802-kube-api-access-wm4k9\") pod \"nmstate-metrics-9b8c8685d-k2x5n\" (UID: \"0d0f7c9a-ffef-4d0b-ae31-0cfd9e864802\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-k2x5n" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.744274 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2476\" (UniqueName: \"kubernetes.io/projected/8813052b-cc40-4fe7-b4de-7aeb3832d4bb-kube-api-access-b2476\") pod \"nmstate-handler-zckpn\" (UID: \"8813052b-cc40-4fe7-b4de-7aeb3832d4bb\") " pod="openshift-nmstate/nmstate-handler-zckpn" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.782316 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-4jmsm"] Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.783142 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-4jmsm" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.785222 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.785887 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-vn4b6" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.785904 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.831807 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-4jmsm"] Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.845099 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/23256f06-964f-4f6f-bedf-6159a92a06d0-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-95n52\" (UID: \"23256f06-964f-4f6f-bedf-6159a92a06d0\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-95n52" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.845180 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/8813052b-cc40-4fe7-b4de-7aeb3832d4bb-ovs-socket\") pod \"nmstate-handler-zckpn\" (UID: \"8813052b-cc40-4fe7-b4de-7aeb3832d4bb\") " pod="openshift-nmstate/nmstate-handler-zckpn" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.845210 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm4k9\" (UniqueName: \"kubernetes.io/projected/0d0f7c9a-ffef-4d0b-ae31-0cfd9e864802-kube-api-access-wm4k9\") pod \"nmstate-metrics-9b8c8685d-k2x5n\" (UID: \"0d0f7c9a-ffef-4d0b-ae31-0cfd9e864802\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-k2x5n" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.845245 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vlmr\" (UniqueName: \"kubernetes.io/projected/89acc01c-091e-402c-a206-8fbc201f0884-kube-api-access-2vlmr\") pod \"nmstate-console-plugin-86f58fcf4-4jmsm\" (UID: \"89acc01c-091e-402c-a206-8fbc201f0884\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-4jmsm" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.845277 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2476\" (UniqueName: \"kubernetes.io/projected/8813052b-cc40-4fe7-b4de-7aeb3832d4bb-kube-api-access-b2476\") pod \"nmstate-handler-zckpn\" (UID: \"8813052b-cc40-4fe7-b4de-7aeb3832d4bb\") " pod="openshift-nmstate/nmstate-handler-zckpn" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.845311 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/89acc01c-091e-402c-a206-8fbc201f0884-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-4jmsm\" (UID: \"89acc01c-091e-402c-a206-8fbc201f0884\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-4jmsm" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.845327 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xbxnm\" (UniqueName: \"kubernetes.io/projected/23256f06-964f-4f6f-bedf-6159a92a06d0-kube-api-access-xbxnm\") pod \"nmstate-webhook-5f558f5558-95n52\" (UID: \"23256f06-964f-4f6f-bedf-6159a92a06d0\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-95n52" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.845368 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/8813052b-cc40-4fe7-b4de-7aeb3832d4bb-dbus-socket\") pod \"nmstate-handler-zckpn\" (UID: \"8813052b-cc40-4fe7-b4de-7aeb3832d4bb\") " pod="openshift-nmstate/nmstate-handler-zckpn" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.845387 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/89acc01c-091e-402c-a206-8fbc201f0884-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-4jmsm\" (UID: \"89acc01c-091e-402c-a206-8fbc201f0884\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-4jmsm" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.845419 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/8813052b-cc40-4fe7-b4de-7aeb3832d4bb-nmstate-lock\") pod \"nmstate-handler-zckpn\" (UID: \"8813052b-cc40-4fe7-b4de-7aeb3832d4bb\") " pod="openshift-nmstate/nmstate-handler-zckpn" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.845478 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/8813052b-cc40-4fe7-b4de-7aeb3832d4bb-nmstate-lock\") pod \"nmstate-handler-zckpn\" (UID: \"8813052b-cc40-4fe7-b4de-7aeb3832d4bb\") " pod="openshift-nmstate/nmstate-handler-zckpn" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.845516 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/8813052b-cc40-4fe7-b4de-7aeb3832d4bb-ovs-socket\") pod \"nmstate-handler-zckpn\" (UID: \"8813052b-cc40-4fe7-b4de-7aeb3832d4bb\") " pod="openshift-nmstate/nmstate-handler-zckpn" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.846075 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/8813052b-cc40-4fe7-b4de-7aeb3832d4bb-dbus-socket\") pod \"nmstate-handler-zckpn\" (UID: \"8813052b-cc40-4fe7-b4de-7aeb3832d4bb\") " pod="openshift-nmstate/nmstate-handler-zckpn" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.859947 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/23256f06-964f-4f6f-bedf-6159a92a06d0-tls-key-pair\") pod \"nmstate-webhook-5f558f5558-95n52\" (UID: \"23256f06-964f-4f6f-bedf-6159a92a06d0\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-95n52" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.860904 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm4k9\" (UniqueName: \"kubernetes.io/projected/0d0f7c9a-ffef-4d0b-ae31-0cfd9e864802-kube-api-access-wm4k9\") pod \"nmstate-metrics-9b8c8685d-k2x5n\" (UID: \"0d0f7c9a-ffef-4d0b-ae31-0cfd9e864802\") " pod="openshift-nmstate/nmstate-metrics-9b8c8685d-k2x5n" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.861371 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xbxnm\" (UniqueName: \"kubernetes.io/projected/23256f06-964f-4f6f-bedf-6159a92a06d0-kube-api-access-xbxnm\") pod \"nmstate-webhook-5f558f5558-95n52\" (UID: \"23256f06-964f-4f6f-bedf-6159a92a06d0\") " pod="openshift-nmstate/nmstate-webhook-5f558f5558-95n52" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.863809 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2476\" (UniqueName: \"kubernetes.io/projected/8813052b-cc40-4fe7-b4de-7aeb3832d4bb-kube-api-access-b2476\") pod \"nmstate-handler-zckpn\" (UID: \"8813052b-cc40-4fe7-b4de-7aeb3832d4bb\") " pod="openshift-nmstate/nmstate-handler-zckpn" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.946304 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/89acc01c-091e-402c-a206-8fbc201f0884-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-4jmsm\" (UID: \"89acc01c-091e-402c-a206-8fbc201f0884\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-4jmsm" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.946414 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vlmr\" (UniqueName: \"kubernetes.io/projected/89acc01c-091e-402c-a206-8fbc201f0884-kube-api-access-2vlmr\") pod \"nmstate-console-plugin-86f58fcf4-4jmsm\" (UID: \"89acc01c-091e-402c-a206-8fbc201f0884\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-4jmsm" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.946458 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/89acc01c-091e-402c-a206-8fbc201f0884-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-4jmsm\" (UID: \"89acc01c-091e-402c-a206-8fbc201f0884\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-4jmsm" Mar 17 09:24:52 crc kubenswrapper[4813]: E0317 09:24:52.946534 4813 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Mar 17 09:24:52 crc kubenswrapper[4813]: E0317 09:24:52.946636 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/89acc01c-091e-402c-a206-8fbc201f0884-plugin-serving-cert podName:89acc01c-091e-402c-a206-8fbc201f0884 nodeName:}" failed. No retries permitted until 2026-03-17 09:24:53.446615294 +0000 UTC m=+915.547418803 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/89acc01c-091e-402c-a206-8fbc201f0884-plugin-serving-cert") pod "nmstate-console-plugin-86f58fcf4-4jmsm" (UID: "89acc01c-091e-402c-a206-8fbc201f0884") : secret "plugin-serving-cert" not found Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.947532 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/89acc01c-091e-402c-a206-8fbc201f0884-nginx-conf\") pod \"nmstate-console-plugin-86f58fcf4-4jmsm\" (UID: \"89acc01c-091e-402c-a206-8fbc201f0884\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-4jmsm" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.960705 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-75685b4cc-h5mtg"] Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.961476 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.965292 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f558f5558-95n52" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.975213 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-k2x5n" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.980878 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-75685b4cc-h5mtg"] Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.995334 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vlmr\" (UniqueName: \"kubernetes.io/projected/89acc01c-091e-402c-a206-8fbc201f0884-kube-api-access-2vlmr\") pod \"nmstate-console-plugin-86f58fcf4-4jmsm\" (UID: \"89acc01c-091e-402c-a206-8fbc201f0884\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-4jmsm" Mar 17 09:24:52 crc kubenswrapper[4813]: I0317 09:24:52.996769 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-zckpn" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.047971 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002-console-serving-cert\") pod \"console-75685b4cc-h5mtg\" (UID: \"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002\") " pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.048009 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002-oauth-serving-cert\") pod \"console-75685b4cc-h5mtg\" (UID: \"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002\") " pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.048064 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002-service-ca\") pod \"console-75685b4cc-h5mtg\" (UID: \"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002\") " pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.048082 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002-trusted-ca-bundle\") pod \"console-75685b4cc-h5mtg\" (UID: \"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002\") " pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.048343 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002-console-config\") pod \"console-75685b4cc-h5mtg\" (UID: \"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002\") " pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.048797 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002-console-oauth-config\") pod \"console-75685b4cc-h5mtg\" (UID: \"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002\") " pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.049469 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cwbq\" (UniqueName: \"kubernetes.io/projected/a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002-kube-api-access-4cwbq\") pod \"console-75685b4cc-h5mtg\" (UID: \"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002\") " pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.152216 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002-console-config\") pod \"console-75685b4cc-h5mtg\" (UID: \"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002\") " pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.152268 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002-console-oauth-config\") pod \"console-75685b4cc-h5mtg\" (UID: \"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002\") " pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.152300 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cwbq\" (UniqueName: \"kubernetes.io/projected/a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002-kube-api-access-4cwbq\") pod \"console-75685b4cc-h5mtg\" (UID: \"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002\") " pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.152336 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002-console-serving-cert\") pod \"console-75685b4cc-h5mtg\" (UID: \"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002\") " pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.152358 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002-oauth-serving-cert\") pod \"console-75685b4cc-h5mtg\" (UID: \"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002\") " pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.152398 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002-service-ca\") pod \"console-75685b4cc-h5mtg\" (UID: \"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002\") " pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.152417 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002-trusted-ca-bundle\") pod \"console-75685b4cc-h5mtg\" (UID: \"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002\") " pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.153143 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002-console-config\") pod \"console-75685b4cc-h5mtg\" (UID: \"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002\") " pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.153281 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002-trusted-ca-bundle\") pod \"console-75685b4cc-h5mtg\" (UID: \"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002\") " pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.153380 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002-oauth-serving-cert\") pod \"console-75685b4cc-h5mtg\" (UID: \"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002\") " pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.154119 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002-service-ca\") pod \"console-75685b4cc-h5mtg\" (UID: \"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002\") " pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.157829 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002-console-oauth-config\") pod \"console-75685b4cc-h5mtg\" (UID: \"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002\") " pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.158117 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002-console-serving-cert\") pod \"console-75685b4cc-h5mtg\" (UID: \"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002\") " pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.167900 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cwbq\" (UniqueName: \"kubernetes.io/projected/a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002-kube-api-access-4cwbq\") pod \"console-75685b4cc-h5mtg\" (UID: \"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002\") " pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.211960 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f558f5558-95n52"] Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.273860 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.417014 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-zckpn" event={"ID":"8813052b-cc40-4fe7-b4de-7aeb3832d4bb","Type":"ContainerStarted","Data":"0ccbd8e297a11cbe591dd00996d53fdc0d630fec44d5f41335f97ac6055deaba"} Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.417694 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-95n52" event={"ID":"23256f06-964f-4f6f-bedf-6159a92a06d0","Type":"ContainerStarted","Data":"6dc3fe13232beb9df80844bb535dc76cb143e117be5f87be6ae5dd6213399f3a"} Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.456941 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/89acc01c-091e-402c-a206-8fbc201f0884-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-4jmsm\" (UID: \"89acc01c-091e-402c-a206-8fbc201f0884\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-4jmsm" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.461044 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/89acc01c-091e-402c-a206-8fbc201f0884-plugin-serving-cert\") pod \"nmstate-console-plugin-86f58fcf4-4jmsm\" (UID: \"89acc01c-091e-402c-a206-8fbc201f0884\") " pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-4jmsm" Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.468959 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-75685b4cc-h5mtg"] Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.473653 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-9b8c8685d-k2x5n"] Mar 17 09:24:53 crc kubenswrapper[4813]: W0317 09:24:53.477714 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0d0f7c9a_ffef_4d0b_ae31_0cfd9e864802.slice/crio-c0a6b2e65df163fbef261de5a87a2a7bab6360a9b92205a98f14856ec1bb78b3 WatchSource:0}: Error finding container c0a6b2e65df163fbef261de5a87a2a7bab6360a9b92205a98f14856ec1bb78b3: Status 404 returned error can't find the container with id c0a6b2e65df163fbef261de5a87a2a7bab6360a9b92205a98f14856ec1bb78b3 Mar 17 09:24:53 crc kubenswrapper[4813]: I0317 09:24:53.695699 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-4jmsm" Mar 17 09:24:54 crc kubenswrapper[4813]: I0317 09:24:54.131930 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-86f58fcf4-4jmsm"] Mar 17 09:24:54 crc kubenswrapper[4813]: W0317 09:24:54.138466 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod89acc01c_091e_402c_a206_8fbc201f0884.slice/crio-3bde01f9f3c75f4065d8accbdee6627ad0dc2b9df28ccb3d8bcf3caba93f3704 WatchSource:0}: Error finding container 3bde01f9f3c75f4065d8accbdee6627ad0dc2b9df28ccb3d8bcf3caba93f3704: Status 404 returned error can't find the container with id 3bde01f9f3c75f4065d8accbdee6627ad0dc2b9df28ccb3d8bcf3caba93f3704 Mar 17 09:24:54 crc kubenswrapper[4813]: I0317 09:24:54.427936 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-4jmsm" event={"ID":"89acc01c-091e-402c-a206-8fbc201f0884","Type":"ContainerStarted","Data":"3bde01f9f3c75f4065d8accbdee6627ad0dc2b9df28ccb3d8bcf3caba93f3704"} Mar 17 09:24:54 crc kubenswrapper[4813]: I0317 09:24:54.429399 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-k2x5n" event={"ID":"0d0f7c9a-ffef-4d0b-ae31-0cfd9e864802","Type":"ContainerStarted","Data":"c0a6b2e65df163fbef261de5a87a2a7bab6360a9b92205a98f14856ec1bb78b3"} Mar 17 09:24:54 crc kubenswrapper[4813]: I0317 09:24:54.432115 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-75685b4cc-h5mtg" event={"ID":"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002","Type":"ContainerStarted","Data":"c9972e6438a582622dc34bd2209cd5d663a340ab86dd36d398e3f6805eac4d32"} Mar 17 09:24:54 crc kubenswrapper[4813]: I0317 09:24:54.432272 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-75685b4cc-h5mtg" event={"ID":"a4c27ea8-ec6e-43b2-8dc8-0a81af0e3002","Type":"ContainerStarted","Data":"5d8d553546b82aa0dbb8e985622435ac9ba22b97cc1401b07d695c5e6f831b93"} Mar 17 09:24:54 crc kubenswrapper[4813]: I0317 09:24:54.455471 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-75685b4cc-h5mtg" podStartSLOduration=2.455454829 podStartE2EDuration="2.455454829s" podCreationTimestamp="2026-03-17 09:24:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:24:54.452474474 +0000 UTC m=+916.553277993" watchObservedRunningTime="2026-03-17 09:24:54.455454829 +0000 UTC m=+916.556258328" Mar 17 09:24:55 crc kubenswrapper[4813]: I0317 09:24:55.946444 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vf8rq" Mar 17 09:24:55 crc kubenswrapper[4813]: I0317 09:24:55.947664 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vf8rq" Mar 17 09:24:56 crc kubenswrapper[4813]: I0317 09:24:56.003009 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vf8rq" Mar 17 09:24:56 crc kubenswrapper[4813]: I0317 09:24:56.447663 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-k2x5n" event={"ID":"0d0f7c9a-ffef-4d0b-ae31-0cfd9e864802","Type":"ContainerStarted","Data":"4e44299800870393d82bdab727321f7241c30a01c67ec7a095b3206bd9106e28"} Mar 17 09:24:56 crc kubenswrapper[4813]: I0317 09:24:56.449651 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f558f5558-95n52" event={"ID":"23256f06-964f-4f6f-bedf-6159a92a06d0","Type":"ContainerStarted","Data":"290bfaa81b7a56e02a0ea38826581c942225e31427953cba876d4eec1ffadb10"} Mar 17 09:24:56 crc kubenswrapper[4813]: I0317 09:24:56.450132 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f558f5558-95n52" Mar 17 09:24:56 crc kubenswrapper[4813]: I0317 09:24:56.452318 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-zckpn" event={"ID":"8813052b-cc40-4fe7-b4de-7aeb3832d4bb","Type":"ContainerStarted","Data":"1aed85f565438ac5d8af326fd89690bbe71638a4c439ee2d9c1c7869e5c3241e"} Mar 17 09:24:56 crc kubenswrapper[4813]: I0317 09:24:56.452358 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-zckpn" Mar 17 09:24:56 crc kubenswrapper[4813]: I0317 09:24:56.474666 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f558f5558-95n52" podStartSLOduration=2.1133361 podStartE2EDuration="4.474635373s" podCreationTimestamp="2026-03-17 09:24:52 +0000 UTC" firstStartedPulling="2026-03-17 09:24:53.217717771 +0000 UTC m=+915.318521270" lastFinishedPulling="2026-03-17 09:24:55.579017014 +0000 UTC m=+917.679820543" observedRunningTime="2026-03-17 09:24:56.467049222 +0000 UTC m=+918.567852731" watchObservedRunningTime="2026-03-17 09:24:56.474635373 +0000 UTC m=+918.575438912" Mar 17 09:24:56 crc kubenswrapper[4813]: I0317 09:24:56.497271 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-zckpn" podStartSLOduration=1.950568952 podStartE2EDuration="4.497202369s" podCreationTimestamp="2026-03-17 09:24:52 +0000 UTC" firstStartedPulling="2026-03-17 09:24:53.031313283 +0000 UTC m=+915.132116782" lastFinishedPulling="2026-03-17 09:24:55.57794666 +0000 UTC m=+917.678750199" observedRunningTime="2026-03-17 09:24:56.484001781 +0000 UTC m=+918.584805320" watchObservedRunningTime="2026-03-17 09:24:56.497202369 +0000 UTC m=+918.598005908" Mar 17 09:24:56 crc kubenswrapper[4813]: I0317 09:24:56.534667 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vf8rq" Mar 17 09:24:56 crc kubenswrapper[4813]: I0317 09:24:56.614090 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vf8rq"] Mar 17 09:24:56 crc kubenswrapper[4813]: I0317 09:24:56.657408 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cm7vw"] Mar 17 09:24:56 crc kubenswrapper[4813]: I0317 09:24:56.657689 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-cm7vw" podUID="ba5c8bcc-1783-4889-a504-4233deca14dd" containerName="registry-server" containerID="cri-o://3b53c27ab24c90a649a1c219348d70d9599a4fdb36136d288a39b0ac47e28c8b" gracePeriod=2 Mar 17 09:24:56 crc kubenswrapper[4813]: I0317 09:24:56.707791 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-marketplace/redhat-operators-cm7vw" podUID="ba5c8bcc-1783-4889-a504-4233deca14dd" containerName="registry-server" probeResult="failure" output="" Mar 17 09:24:56 crc kubenswrapper[4813]: I0317 09:24:56.707804 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/redhat-operators-cm7vw" podUID="ba5c8bcc-1783-4889-a504-4233deca14dd" containerName="registry-server" probeResult="failure" output="" Mar 17 09:24:57 crc kubenswrapper[4813]: I0317 09:24:57.890726 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cm7vw" Mar 17 09:24:57 crc kubenswrapper[4813]: I0317 09:24:57.907069 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba5c8bcc-1783-4889-a504-4233deca14dd-catalog-content\") pod \"ba5c8bcc-1783-4889-a504-4233deca14dd\" (UID: \"ba5c8bcc-1783-4889-a504-4233deca14dd\") " Mar 17 09:24:57 crc kubenswrapper[4813]: I0317 09:24:57.907112 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sffwb\" (UniqueName: \"kubernetes.io/projected/ba5c8bcc-1783-4889-a504-4233deca14dd-kube-api-access-sffwb\") pod \"ba5c8bcc-1783-4889-a504-4233deca14dd\" (UID: \"ba5c8bcc-1783-4889-a504-4233deca14dd\") " Mar 17 09:24:57 crc kubenswrapper[4813]: I0317 09:24:57.907222 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba5c8bcc-1783-4889-a504-4233deca14dd-utilities\") pod \"ba5c8bcc-1783-4889-a504-4233deca14dd\" (UID: \"ba5c8bcc-1783-4889-a504-4233deca14dd\") " Mar 17 09:24:57 crc kubenswrapper[4813]: I0317 09:24:57.908088 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba5c8bcc-1783-4889-a504-4233deca14dd-utilities" (OuterVolumeSpecName: "utilities") pod "ba5c8bcc-1783-4889-a504-4233deca14dd" (UID: "ba5c8bcc-1783-4889-a504-4233deca14dd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:24:57 crc kubenswrapper[4813]: I0317 09:24:57.926262 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba5c8bcc-1783-4889-a504-4233deca14dd-kube-api-access-sffwb" (OuterVolumeSpecName: "kube-api-access-sffwb") pod "ba5c8bcc-1783-4889-a504-4233deca14dd" (UID: "ba5c8bcc-1783-4889-a504-4233deca14dd"). InnerVolumeSpecName "kube-api-access-sffwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:24:58 crc kubenswrapper[4813]: I0317 09:24:58.009006 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba5c8bcc-1783-4889-a504-4233deca14dd-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:24:58 crc kubenswrapper[4813]: I0317 09:24:58.009492 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sffwb\" (UniqueName: \"kubernetes.io/projected/ba5c8bcc-1783-4889-a504-4233deca14dd-kube-api-access-sffwb\") on node \"crc\" DevicePath \"\"" Mar 17 09:24:58 crc kubenswrapper[4813]: I0317 09:24:58.070896 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba5c8bcc-1783-4889-a504-4233deca14dd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba5c8bcc-1783-4889-a504-4233deca14dd" (UID: "ba5c8bcc-1783-4889-a504-4233deca14dd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:24:58 crc kubenswrapper[4813]: I0317 09:24:58.110866 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba5c8bcc-1783-4889-a504-4233deca14dd-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:24:58 crc kubenswrapper[4813]: I0317 09:24:58.477457 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-4jmsm" event={"ID":"89acc01c-091e-402c-a206-8fbc201f0884","Type":"ContainerStarted","Data":"2bc3db5102016973c4d5a77d9e72820204b47e0a5409e1bd7d804321d6e57a7d"} Mar 17 09:24:58 crc kubenswrapper[4813]: I0317 09:24:58.482919 4813 generic.go:334] "Generic (PLEG): container finished" podID="ba5c8bcc-1783-4889-a504-4233deca14dd" containerID="3b53c27ab24c90a649a1c219348d70d9599a4fdb36136d288a39b0ac47e28c8b" exitCode=0 Mar 17 09:24:58 crc kubenswrapper[4813]: I0317 09:24:58.483436 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-cm7vw" Mar 17 09:24:58 crc kubenswrapper[4813]: I0317 09:24:58.483429 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cm7vw" event={"ID":"ba5c8bcc-1783-4889-a504-4233deca14dd","Type":"ContainerDied","Data":"3b53c27ab24c90a649a1c219348d70d9599a4fdb36136d288a39b0ac47e28c8b"} Mar 17 09:24:58 crc kubenswrapper[4813]: I0317 09:24:58.483572 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-cm7vw" event={"ID":"ba5c8bcc-1783-4889-a504-4233deca14dd","Type":"ContainerDied","Data":"cec77bbf462499140275b2ada28ce2e75b9abd072218f04f19842ae47a50f941"} Mar 17 09:24:58 crc kubenswrapper[4813]: I0317 09:24:58.483628 4813 scope.go:117] "RemoveContainer" containerID="3b53c27ab24c90a649a1c219348d70d9599a4fdb36136d288a39b0ac47e28c8b" Mar 17 09:24:58 crc kubenswrapper[4813]: I0317 09:24:58.496369 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-86f58fcf4-4jmsm" podStartSLOduration=2.923968391 podStartE2EDuration="6.496268625s" podCreationTimestamp="2026-03-17 09:24:52 +0000 UTC" firstStartedPulling="2026-03-17 09:24:54.140417029 +0000 UTC m=+916.241220528" lastFinishedPulling="2026-03-17 09:24:57.712717263 +0000 UTC m=+919.813520762" observedRunningTime="2026-03-17 09:24:58.493244399 +0000 UTC m=+920.594047898" watchObservedRunningTime="2026-03-17 09:24:58.496268625 +0000 UTC m=+920.597072124" Mar 17 09:24:58 crc kubenswrapper[4813]: I0317 09:24:58.552448 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-cm7vw"] Mar 17 09:24:58 crc kubenswrapper[4813]: I0317 09:24:58.557028 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-cm7vw"] Mar 17 09:24:58 crc kubenswrapper[4813]: I0317 09:24:58.737854 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba5c8bcc-1783-4889-a504-4233deca14dd" path="/var/lib/kubelet/pods/ba5c8bcc-1783-4889-a504-4233deca14dd/volumes" Mar 17 09:24:59 crc kubenswrapper[4813]: I0317 09:24:59.165419 4813 scope.go:117] "RemoveContainer" containerID="bc81c19ce8a9d9e4f3ab3c76d4baf2d0f737dc5997af0720ec4b525de2db7ccf" Mar 17 09:24:59 crc kubenswrapper[4813]: I0317 09:24:59.218443 4813 scope.go:117] "RemoveContainer" containerID="4acaa380639f15fa272d0ea3f564e8fa0e2a137daa7f0b4d711c533350647f7c" Mar 17 09:24:59 crc kubenswrapper[4813]: I0317 09:24:59.239901 4813 scope.go:117] "RemoveContainer" containerID="3b53c27ab24c90a649a1c219348d70d9599a4fdb36136d288a39b0ac47e28c8b" Mar 17 09:24:59 crc kubenswrapper[4813]: E0317 09:24:59.240320 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3b53c27ab24c90a649a1c219348d70d9599a4fdb36136d288a39b0ac47e28c8b\": container with ID starting with 3b53c27ab24c90a649a1c219348d70d9599a4fdb36136d288a39b0ac47e28c8b not found: ID does not exist" containerID="3b53c27ab24c90a649a1c219348d70d9599a4fdb36136d288a39b0ac47e28c8b" Mar 17 09:24:59 crc kubenswrapper[4813]: I0317 09:24:59.240371 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3b53c27ab24c90a649a1c219348d70d9599a4fdb36136d288a39b0ac47e28c8b"} err="failed to get container status \"3b53c27ab24c90a649a1c219348d70d9599a4fdb36136d288a39b0ac47e28c8b\": rpc error: code = NotFound desc = could not find container \"3b53c27ab24c90a649a1c219348d70d9599a4fdb36136d288a39b0ac47e28c8b\": container with ID starting with 3b53c27ab24c90a649a1c219348d70d9599a4fdb36136d288a39b0ac47e28c8b not found: ID does not exist" Mar 17 09:24:59 crc kubenswrapper[4813]: I0317 09:24:59.240405 4813 scope.go:117] "RemoveContainer" containerID="bc81c19ce8a9d9e4f3ab3c76d4baf2d0f737dc5997af0720ec4b525de2db7ccf" Mar 17 09:24:59 crc kubenswrapper[4813]: E0317 09:24:59.240643 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc81c19ce8a9d9e4f3ab3c76d4baf2d0f737dc5997af0720ec4b525de2db7ccf\": container with ID starting with bc81c19ce8a9d9e4f3ab3c76d4baf2d0f737dc5997af0720ec4b525de2db7ccf not found: ID does not exist" containerID="bc81c19ce8a9d9e4f3ab3c76d4baf2d0f737dc5997af0720ec4b525de2db7ccf" Mar 17 09:24:59 crc kubenswrapper[4813]: I0317 09:24:59.240671 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc81c19ce8a9d9e4f3ab3c76d4baf2d0f737dc5997af0720ec4b525de2db7ccf"} err="failed to get container status \"bc81c19ce8a9d9e4f3ab3c76d4baf2d0f737dc5997af0720ec4b525de2db7ccf\": rpc error: code = NotFound desc = could not find container \"bc81c19ce8a9d9e4f3ab3c76d4baf2d0f737dc5997af0720ec4b525de2db7ccf\": container with ID starting with bc81c19ce8a9d9e4f3ab3c76d4baf2d0f737dc5997af0720ec4b525de2db7ccf not found: ID does not exist" Mar 17 09:24:59 crc kubenswrapper[4813]: I0317 09:24:59.240692 4813 scope.go:117] "RemoveContainer" containerID="4acaa380639f15fa272d0ea3f564e8fa0e2a137daa7f0b4d711c533350647f7c" Mar 17 09:24:59 crc kubenswrapper[4813]: E0317 09:24:59.241045 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4acaa380639f15fa272d0ea3f564e8fa0e2a137daa7f0b4d711c533350647f7c\": container with ID starting with 4acaa380639f15fa272d0ea3f564e8fa0e2a137daa7f0b4d711c533350647f7c not found: ID does not exist" containerID="4acaa380639f15fa272d0ea3f564e8fa0e2a137daa7f0b4d711c533350647f7c" Mar 17 09:24:59 crc kubenswrapper[4813]: I0317 09:24:59.241084 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4acaa380639f15fa272d0ea3f564e8fa0e2a137daa7f0b4d711c533350647f7c"} err="failed to get container status \"4acaa380639f15fa272d0ea3f564e8fa0e2a137daa7f0b4d711c533350647f7c\": rpc error: code = NotFound desc = could not find container \"4acaa380639f15fa272d0ea3f564e8fa0e2a137daa7f0b4d711c533350647f7c\": container with ID starting with 4acaa380639f15fa272d0ea3f564e8fa0e2a137daa7f0b4d711c533350647f7c not found: ID does not exist" Mar 17 09:24:59 crc kubenswrapper[4813]: I0317 09:24:59.489386 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-k2x5n" event={"ID":"0d0f7c9a-ffef-4d0b-ae31-0cfd9e864802","Type":"ContainerStarted","Data":"f76cd93fe5eb787e1a5a878d2de6aa5e78cf274f7394d388756f243c00381e4b"} Mar 17 09:24:59 crc kubenswrapper[4813]: I0317 09:24:59.515587 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-9b8c8685d-k2x5n" podStartSLOduration=1.755462751 podStartE2EDuration="7.51556631s" podCreationTimestamp="2026-03-17 09:24:52 +0000 UTC" firstStartedPulling="2026-03-17 09:24:53.480208233 +0000 UTC m=+915.581011732" lastFinishedPulling="2026-03-17 09:24:59.240311792 +0000 UTC m=+921.341115291" observedRunningTime="2026-03-17 09:24:59.514717343 +0000 UTC m=+921.615520862" watchObservedRunningTime="2026-03-17 09:24:59.51556631 +0000 UTC m=+921.616369809" Mar 17 09:25:03 crc kubenswrapper[4813]: I0317 09:25:03.036810 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-zckpn" Mar 17 09:25:03 crc kubenswrapper[4813]: I0317 09:25:03.274409 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:25:03 crc kubenswrapper[4813]: I0317 09:25:03.274492 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:25:03 crc kubenswrapper[4813]: I0317 09:25:03.282739 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:25:03 crc kubenswrapper[4813]: I0317 09:25:03.533206 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-75685b4cc-h5mtg" Mar 17 09:25:03 crc kubenswrapper[4813]: I0317 09:25:03.617367 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-2w9vb"] Mar 17 09:25:12 crc kubenswrapper[4813]: I0317 09:25:12.974928 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f558f5558-95n52" Mar 17 09:25:27 crc kubenswrapper[4813]: I0317 09:25:27.684590 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98"] Mar 17 09:25:27 crc kubenswrapper[4813]: E0317 09:25:27.685347 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba5c8bcc-1783-4889-a504-4233deca14dd" containerName="extract-utilities" Mar 17 09:25:27 crc kubenswrapper[4813]: I0317 09:25:27.685362 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba5c8bcc-1783-4889-a504-4233deca14dd" containerName="extract-utilities" Mar 17 09:25:27 crc kubenswrapper[4813]: E0317 09:25:27.685373 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba5c8bcc-1783-4889-a504-4233deca14dd" containerName="extract-content" Mar 17 09:25:27 crc kubenswrapper[4813]: I0317 09:25:27.685381 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba5c8bcc-1783-4889-a504-4233deca14dd" containerName="extract-content" Mar 17 09:25:27 crc kubenswrapper[4813]: E0317 09:25:27.685408 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba5c8bcc-1783-4889-a504-4233deca14dd" containerName="registry-server" Mar 17 09:25:27 crc kubenswrapper[4813]: I0317 09:25:27.685416 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba5c8bcc-1783-4889-a504-4233deca14dd" containerName="registry-server" Mar 17 09:25:27 crc kubenswrapper[4813]: I0317 09:25:27.685719 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba5c8bcc-1783-4889-a504-4233deca14dd" containerName="registry-server" Mar 17 09:25:27 crc kubenswrapper[4813]: I0317 09:25:27.686469 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98" Mar 17 09:25:27 crc kubenswrapper[4813]: I0317 09:25:27.688849 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 17 09:25:27 crc kubenswrapper[4813]: I0317 09:25:27.705944 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98"] Mar 17 09:25:27 crc kubenswrapper[4813]: I0317 09:25:27.755044 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5m6b7\" (UniqueName: \"kubernetes.io/projected/0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2-kube-api-access-5m6b7\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98\" (UID: \"0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98" Mar 17 09:25:27 crc kubenswrapper[4813]: I0317 09:25:27.755204 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98\" (UID: \"0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98" Mar 17 09:25:27 crc kubenswrapper[4813]: I0317 09:25:27.755305 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98\" (UID: \"0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98" Mar 17 09:25:27 crc kubenswrapper[4813]: I0317 09:25:27.856461 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98\" (UID: \"0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98" Mar 17 09:25:27 crc kubenswrapper[4813]: I0317 09:25:27.856546 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5m6b7\" (UniqueName: \"kubernetes.io/projected/0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2-kube-api-access-5m6b7\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98\" (UID: \"0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98" Mar 17 09:25:27 crc kubenswrapper[4813]: I0317 09:25:27.856690 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98\" (UID: \"0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98" Mar 17 09:25:27 crc kubenswrapper[4813]: I0317 09:25:27.857564 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2-util\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98\" (UID: \"0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98" Mar 17 09:25:27 crc kubenswrapper[4813]: I0317 09:25:27.857681 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2-bundle\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98\" (UID: \"0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98" Mar 17 09:25:27 crc kubenswrapper[4813]: I0317 09:25:27.880910 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5m6b7\" (UniqueName: \"kubernetes.io/projected/0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2-kube-api-access-5m6b7\") pod \"2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98\" (UID: \"0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2\") " pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98" Mar 17 09:25:28 crc kubenswrapper[4813]: I0317 09:25:28.003267 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98" Mar 17 09:25:28 crc kubenswrapper[4813]: I0317 09:25:28.217502 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98"] Mar 17 09:25:28 crc kubenswrapper[4813]: W0317 09:25:28.225310 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b47f2fe_3d7e_4550_a89f_79d3af2cc8f2.slice/crio-1e6311fe5fd8b986427f8b8c87af55cecb9adb7b3088f893b41b37c139b32523 WatchSource:0}: Error finding container 1e6311fe5fd8b986427f8b8c87af55cecb9adb7b3088f893b41b37c139b32523: Status 404 returned error can't find the container with id 1e6311fe5fd8b986427f8b8c87af55cecb9adb7b3088f893b41b37c139b32523 Mar 17 09:25:28 crc kubenswrapper[4813]: I0317 09:25:28.686285 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-2w9vb" podUID="f7603094-bb1d-4d77-b463-2270003b2805" containerName="console" containerID="cri-o://011538bd3ac194eac4eaac10f92db465d10f8ff49e5ec1b4f73b7400a510e831" gracePeriod=15 Mar 17 09:25:28 crc kubenswrapper[4813]: I0317 09:25:28.734989 4813 generic.go:334] "Generic (PLEG): container finished" podID="0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2" containerID="3b4a7b9726de6d887c43e1868bf38725c681afdf086c5030a01ddd55c464c7d1" exitCode=0 Mar 17 09:25:28 crc kubenswrapper[4813]: I0317 09:25:28.743268 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 09:25:28 crc kubenswrapper[4813]: I0317 09:25:28.743760 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98" event={"ID":"0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2","Type":"ContainerDied","Data":"3b4a7b9726de6d887c43e1868bf38725c681afdf086c5030a01ddd55c464c7d1"} Mar 17 09:25:28 crc kubenswrapper[4813]: I0317 09:25:28.743810 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98" event={"ID":"0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2","Type":"ContainerStarted","Data":"1e6311fe5fd8b986427f8b8c87af55cecb9adb7b3088f893b41b37c139b32523"} Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.165227 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-2w9vb_f7603094-bb1d-4d77-b463-2270003b2805/console/0.log" Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.165519 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.277769 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f7603094-bb1d-4d77-b463-2270003b2805-console-oauth-config\") pod \"f7603094-bb1d-4d77-b463-2270003b2805\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.277861 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-console-config\") pod \"f7603094-bb1d-4d77-b463-2270003b2805\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.277905 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-oauth-serving-cert\") pod \"f7603094-bb1d-4d77-b463-2270003b2805\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.277976 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-service-ca\") pod \"f7603094-bb1d-4d77-b463-2270003b2805\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.277992 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-trusted-ca-bundle\") pod \"f7603094-bb1d-4d77-b463-2270003b2805\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.278025 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f7603094-bb1d-4d77-b463-2270003b2805-console-serving-cert\") pod \"f7603094-bb1d-4d77-b463-2270003b2805\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.278046 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngkkm\" (UniqueName: \"kubernetes.io/projected/f7603094-bb1d-4d77-b463-2270003b2805-kube-api-access-ngkkm\") pod \"f7603094-bb1d-4d77-b463-2270003b2805\" (UID: \"f7603094-bb1d-4d77-b463-2270003b2805\") " Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.278918 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-service-ca" (OuterVolumeSpecName: "service-ca") pod "f7603094-bb1d-4d77-b463-2270003b2805" (UID: "f7603094-bb1d-4d77-b463-2270003b2805"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.278886 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "f7603094-bb1d-4d77-b463-2270003b2805" (UID: "f7603094-bb1d-4d77-b463-2270003b2805"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.279380 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "f7603094-bb1d-4d77-b463-2270003b2805" (UID: "f7603094-bb1d-4d77-b463-2270003b2805"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.279785 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-console-config" (OuterVolumeSpecName: "console-config") pod "f7603094-bb1d-4d77-b463-2270003b2805" (UID: "f7603094-bb1d-4d77-b463-2270003b2805"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.283741 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7603094-bb1d-4d77-b463-2270003b2805-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "f7603094-bb1d-4d77-b463-2270003b2805" (UID: "f7603094-bb1d-4d77-b463-2270003b2805"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.283995 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f7603094-bb1d-4d77-b463-2270003b2805-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "f7603094-bb1d-4d77-b463-2270003b2805" (UID: "f7603094-bb1d-4d77-b463-2270003b2805"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.286694 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f7603094-bb1d-4d77-b463-2270003b2805-kube-api-access-ngkkm" (OuterVolumeSpecName: "kube-api-access-ngkkm") pod "f7603094-bb1d-4d77-b463-2270003b2805" (UID: "f7603094-bb1d-4d77-b463-2270003b2805"). InnerVolumeSpecName "kube-api-access-ngkkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.380503 4813 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-console-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.380582 4813 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.380657 4813 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-service-ca\") on node \"crc\" DevicePath \"\"" Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.380681 4813 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f7603094-bb1d-4d77-b463-2270003b2805-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.380705 4813 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/f7603094-bb1d-4d77-b463-2270003b2805-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.380729 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngkkm\" (UniqueName: \"kubernetes.io/projected/f7603094-bb1d-4d77-b463-2270003b2805-kube-api-access-ngkkm\") on node \"crc\" DevicePath \"\"" Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.380755 4813 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/f7603094-bb1d-4d77-b463-2270003b2805-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.746037 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-2w9vb_f7603094-bb1d-4d77-b463-2270003b2805/console/0.log" Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.746111 4813 generic.go:334] "Generic (PLEG): container finished" podID="f7603094-bb1d-4d77-b463-2270003b2805" containerID="011538bd3ac194eac4eaac10f92db465d10f8ff49e5ec1b4f73b7400a510e831" exitCode=2 Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.746154 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2w9vb" event={"ID":"f7603094-bb1d-4d77-b463-2270003b2805","Type":"ContainerDied","Data":"011538bd3ac194eac4eaac10f92db465d10f8ff49e5ec1b4f73b7400a510e831"} Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.746192 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-2w9vb" event={"ID":"f7603094-bb1d-4d77-b463-2270003b2805","Type":"ContainerDied","Data":"2967cb5bd44fd0526a4158a5226c69df2a26093d950f605e61c3ba21d8a5f068"} Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.746219 4813 scope.go:117] "RemoveContainer" containerID="011538bd3ac194eac4eaac10f92db465d10f8ff49e5ec1b4f73b7400a510e831" Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.746384 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-2w9vb" Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.774825 4813 scope.go:117] "RemoveContainer" containerID="011538bd3ac194eac4eaac10f92db465d10f8ff49e5ec1b4f73b7400a510e831" Mar 17 09:25:29 crc kubenswrapper[4813]: E0317 09:25:29.775740 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"011538bd3ac194eac4eaac10f92db465d10f8ff49e5ec1b4f73b7400a510e831\": container with ID starting with 011538bd3ac194eac4eaac10f92db465d10f8ff49e5ec1b4f73b7400a510e831 not found: ID does not exist" containerID="011538bd3ac194eac4eaac10f92db465d10f8ff49e5ec1b4f73b7400a510e831" Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.775807 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"011538bd3ac194eac4eaac10f92db465d10f8ff49e5ec1b4f73b7400a510e831"} err="failed to get container status \"011538bd3ac194eac4eaac10f92db465d10f8ff49e5ec1b4f73b7400a510e831\": rpc error: code = NotFound desc = could not find container \"011538bd3ac194eac4eaac10f92db465d10f8ff49e5ec1b4f73b7400a510e831\": container with ID starting with 011538bd3ac194eac4eaac10f92db465d10f8ff49e5ec1b4f73b7400a510e831 not found: ID does not exist" Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.811214 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-2w9vb"] Mar 17 09:25:29 crc kubenswrapper[4813]: I0317 09:25:29.816140 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-2w9vb"] Mar 17 09:25:30 crc kubenswrapper[4813]: I0317 09:25:30.742273 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f7603094-bb1d-4d77-b463-2270003b2805" path="/var/lib/kubelet/pods/f7603094-bb1d-4d77-b463-2270003b2805/volumes" Mar 17 09:25:30 crc kubenswrapper[4813]: I0317 09:25:30.756492 4813 generic.go:334] "Generic (PLEG): container finished" podID="0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2" containerID="816112a2e1948cc0dc1fd99c06700d8fa0c6de978404a1de5cfbd1a40c196957" exitCode=0 Mar 17 09:25:30 crc kubenswrapper[4813]: I0317 09:25:30.756566 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98" event={"ID":"0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2","Type":"ContainerDied","Data":"816112a2e1948cc0dc1fd99c06700d8fa0c6de978404a1de5cfbd1a40c196957"} Mar 17 09:25:31 crc kubenswrapper[4813]: I0317 09:25:31.770072 4813 generic.go:334] "Generic (PLEG): container finished" podID="0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2" containerID="7893db3dd11d157fd846046ddfd777f651bbe135b7ab641c1821bfe7cc82a5ee" exitCode=0 Mar 17 09:25:31 crc kubenswrapper[4813]: I0317 09:25:31.770156 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98" event={"ID":"0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2","Type":"ContainerDied","Data":"7893db3dd11d157fd846046ddfd777f651bbe135b7ab641c1821bfe7cc82a5ee"} Mar 17 09:25:33 crc kubenswrapper[4813]: I0317 09:25:33.098441 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98" Mar 17 09:25:33 crc kubenswrapper[4813]: I0317 09:25:33.240137 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2-bundle\") pod \"0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2\" (UID: \"0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2\") " Mar 17 09:25:33 crc kubenswrapper[4813]: I0317 09:25:33.240546 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2-util\") pod \"0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2\" (UID: \"0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2\") " Mar 17 09:25:33 crc kubenswrapper[4813]: I0317 09:25:33.240945 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5m6b7\" (UniqueName: \"kubernetes.io/projected/0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2-kube-api-access-5m6b7\") pod \"0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2\" (UID: \"0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2\") " Mar 17 09:25:33 crc kubenswrapper[4813]: I0317 09:25:33.243188 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2-bundle" (OuterVolumeSpecName: "bundle") pod "0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2" (UID: "0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:25:33 crc kubenswrapper[4813]: I0317 09:25:33.248523 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2-kube-api-access-5m6b7" (OuterVolumeSpecName: "kube-api-access-5m6b7") pod "0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2" (UID: "0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2"). InnerVolumeSpecName "kube-api-access-5m6b7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:25:33 crc kubenswrapper[4813]: I0317 09:25:33.275918 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2-util" (OuterVolumeSpecName: "util") pod "0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2" (UID: "0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:25:33 crc kubenswrapper[4813]: I0317 09:25:33.343281 4813 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:25:33 crc kubenswrapper[4813]: I0317 09:25:33.343333 4813 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2-util\") on node \"crc\" DevicePath \"\"" Mar 17 09:25:33 crc kubenswrapper[4813]: I0317 09:25:33.343387 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5m6b7\" (UniqueName: \"kubernetes.io/projected/0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2-kube-api-access-5m6b7\") on node \"crc\" DevicePath \"\"" Mar 17 09:25:33 crc kubenswrapper[4813]: I0317 09:25:33.791334 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98" event={"ID":"0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2","Type":"ContainerDied","Data":"1e6311fe5fd8b986427f8b8c87af55cecb9adb7b3088f893b41b37c139b32523"} Mar 17 09:25:33 crc kubenswrapper[4813]: I0317 09:25:33.792022 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1e6311fe5fd8b986427f8b8c87af55cecb9adb7b3088f893b41b37c139b32523" Mar 17 09:25:33 crc kubenswrapper[4813]: I0317 09:25:33.791465 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98" Mar 17 09:25:42 crc kubenswrapper[4813]: I0317 09:25:42.877843 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-56666fbc9b-dj55n"] Mar 17 09:25:42 crc kubenswrapper[4813]: E0317 09:25:42.878322 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f7603094-bb1d-4d77-b463-2270003b2805" containerName="console" Mar 17 09:25:42 crc kubenswrapper[4813]: I0317 09:25:42.878337 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f7603094-bb1d-4d77-b463-2270003b2805" containerName="console" Mar 17 09:25:42 crc kubenswrapper[4813]: E0317 09:25:42.878350 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2" containerName="pull" Mar 17 09:25:42 crc kubenswrapper[4813]: I0317 09:25:42.878358 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2" containerName="pull" Mar 17 09:25:42 crc kubenswrapper[4813]: E0317 09:25:42.878374 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2" containerName="extract" Mar 17 09:25:42 crc kubenswrapper[4813]: I0317 09:25:42.878382 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2" containerName="extract" Mar 17 09:25:42 crc kubenswrapper[4813]: E0317 09:25:42.878405 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2" containerName="util" Mar 17 09:25:42 crc kubenswrapper[4813]: I0317 09:25:42.878412 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2" containerName="util" Mar 17 09:25:42 crc kubenswrapper[4813]: I0317 09:25:42.878530 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2" containerName="extract" Mar 17 09:25:42 crc kubenswrapper[4813]: I0317 09:25:42.878542 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f7603094-bb1d-4d77-b463-2270003b2805" containerName="console" Mar 17 09:25:42 crc kubenswrapper[4813]: I0317 09:25:42.878957 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-56666fbc9b-dj55n" Mar 17 09:25:42 crc kubenswrapper[4813]: I0317 09:25:42.882464 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Mar 17 09:25:42 crc kubenswrapper[4813]: I0317 09:25:42.882509 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-5q4gk" Mar 17 09:25:42 crc kubenswrapper[4813]: I0317 09:25:42.882834 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Mar 17 09:25:42 crc kubenswrapper[4813]: I0317 09:25:42.882575 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Mar 17 09:25:42 crc kubenswrapper[4813]: I0317 09:25:42.883497 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Mar 17 09:25:42 crc kubenswrapper[4813]: I0317 09:25:42.903170 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-56666fbc9b-dj55n"] Mar 17 09:25:42 crc kubenswrapper[4813]: I0317 09:25:42.998497 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmr57\" (UniqueName: \"kubernetes.io/projected/a63b7bd8-1a98-40d9-9f56-c7662d229b3d-kube-api-access-dmr57\") pod \"metallb-operator-controller-manager-56666fbc9b-dj55n\" (UID: \"a63b7bd8-1a98-40d9-9f56-c7662d229b3d\") " pod="metallb-system/metallb-operator-controller-manager-56666fbc9b-dj55n" Mar 17 09:25:42 crc kubenswrapper[4813]: I0317 09:25:42.998565 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a63b7bd8-1a98-40d9-9f56-c7662d229b3d-apiservice-cert\") pod \"metallb-operator-controller-manager-56666fbc9b-dj55n\" (UID: \"a63b7bd8-1a98-40d9-9f56-c7662d229b3d\") " pod="metallb-system/metallb-operator-controller-manager-56666fbc9b-dj55n" Mar 17 09:25:42 crc kubenswrapper[4813]: I0317 09:25:42.998673 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a63b7bd8-1a98-40d9-9f56-c7662d229b3d-webhook-cert\") pod \"metallb-operator-controller-manager-56666fbc9b-dj55n\" (UID: \"a63b7bd8-1a98-40d9-9f56-c7662d229b3d\") " pod="metallb-system/metallb-operator-controller-manager-56666fbc9b-dj55n" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.099284 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmr57\" (UniqueName: \"kubernetes.io/projected/a63b7bd8-1a98-40d9-9f56-c7662d229b3d-kube-api-access-dmr57\") pod \"metallb-operator-controller-manager-56666fbc9b-dj55n\" (UID: \"a63b7bd8-1a98-40d9-9f56-c7662d229b3d\") " pod="metallb-system/metallb-operator-controller-manager-56666fbc9b-dj55n" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.099355 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a63b7bd8-1a98-40d9-9f56-c7662d229b3d-apiservice-cert\") pod \"metallb-operator-controller-manager-56666fbc9b-dj55n\" (UID: \"a63b7bd8-1a98-40d9-9f56-c7662d229b3d\") " pod="metallb-system/metallb-operator-controller-manager-56666fbc9b-dj55n" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.099399 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a63b7bd8-1a98-40d9-9f56-c7662d229b3d-webhook-cert\") pod \"metallb-operator-controller-manager-56666fbc9b-dj55n\" (UID: \"a63b7bd8-1a98-40d9-9f56-c7662d229b3d\") " pod="metallb-system/metallb-operator-controller-manager-56666fbc9b-dj55n" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.105099 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/a63b7bd8-1a98-40d9-9f56-c7662d229b3d-apiservice-cert\") pod \"metallb-operator-controller-manager-56666fbc9b-dj55n\" (UID: \"a63b7bd8-1a98-40d9-9f56-c7662d229b3d\") " pod="metallb-system/metallb-operator-controller-manager-56666fbc9b-dj55n" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.109485 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/a63b7bd8-1a98-40d9-9f56-c7662d229b3d-webhook-cert\") pod \"metallb-operator-controller-manager-56666fbc9b-dj55n\" (UID: \"a63b7bd8-1a98-40d9-9f56-c7662d229b3d\") " pod="metallb-system/metallb-operator-controller-manager-56666fbc9b-dj55n" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.121823 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmr57\" (UniqueName: \"kubernetes.io/projected/a63b7bd8-1a98-40d9-9f56-c7662d229b3d-kube-api-access-dmr57\") pod \"metallb-operator-controller-manager-56666fbc9b-dj55n\" (UID: \"a63b7bd8-1a98-40d9-9f56-c7662d229b3d\") " pod="metallb-system/metallb-operator-controller-manager-56666fbc9b-dj55n" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.195917 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-56666fbc9b-dj55n" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.214218 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-55494b4d7-h6xch"] Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.215095 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-55494b4d7-h6xch" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.219924 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.220075 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-rqvhx" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.220305 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.248580 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-55494b4d7-h6xch"] Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.301284 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f83986c8-b147-48b9-84ec-f09d01ddd8ce-webhook-cert\") pod \"metallb-operator-webhook-server-55494b4d7-h6xch\" (UID: \"f83986c8-b147-48b9-84ec-f09d01ddd8ce\") " pod="metallb-system/metallb-operator-webhook-server-55494b4d7-h6xch" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.301626 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f83986c8-b147-48b9-84ec-f09d01ddd8ce-apiservice-cert\") pod \"metallb-operator-webhook-server-55494b4d7-h6xch\" (UID: \"f83986c8-b147-48b9-84ec-f09d01ddd8ce\") " pod="metallb-system/metallb-operator-webhook-server-55494b4d7-h6xch" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.301788 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8ckv\" (UniqueName: \"kubernetes.io/projected/f83986c8-b147-48b9-84ec-f09d01ddd8ce-kube-api-access-r8ckv\") pod \"metallb-operator-webhook-server-55494b4d7-h6xch\" (UID: \"f83986c8-b147-48b9-84ec-f09d01ddd8ce\") " pod="metallb-system/metallb-operator-webhook-server-55494b4d7-h6xch" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.404194 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f83986c8-b147-48b9-84ec-f09d01ddd8ce-webhook-cert\") pod \"metallb-operator-webhook-server-55494b4d7-h6xch\" (UID: \"f83986c8-b147-48b9-84ec-f09d01ddd8ce\") " pod="metallb-system/metallb-operator-webhook-server-55494b4d7-h6xch" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.404401 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f83986c8-b147-48b9-84ec-f09d01ddd8ce-apiservice-cert\") pod \"metallb-operator-webhook-server-55494b4d7-h6xch\" (UID: \"f83986c8-b147-48b9-84ec-f09d01ddd8ce\") " pod="metallb-system/metallb-operator-webhook-server-55494b4d7-h6xch" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.404503 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8ckv\" (UniqueName: \"kubernetes.io/projected/f83986c8-b147-48b9-84ec-f09d01ddd8ce-kube-api-access-r8ckv\") pod \"metallb-operator-webhook-server-55494b4d7-h6xch\" (UID: \"f83986c8-b147-48b9-84ec-f09d01ddd8ce\") " pod="metallb-system/metallb-operator-webhook-server-55494b4d7-h6xch" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.412242 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f83986c8-b147-48b9-84ec-f09d01ddd8ce-webhook-cert\") pod \"metallb-operator-webhook-server-55494b4d7-h6xch\" (UID: \"f83986c8-b147-48b9-84ec-f09d01ddd8ce\") " pod="metallb-system/metallb-operator-webhook-server-55494b4d7-h6xch" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.412921 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f83986c8-b147-48b9-84ec-f09d01ddd8ce-apiservice-cert\") pod \"metallb-operator-webhook-server-55494b4d7-h6xch\" (UID: \"f83986c8-b147-48b9-84ec-f09d01ddd8ce\") " pod="metallb-system/metallb-operator-webhook-server-55494b4d7-h6xch" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.440105 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8ckv\" (UniqueName: \"kubernetes.io/projected/f83986c8-b147-48b9-84ec-f09d01ddd8ce-kube-api-access-r8ckv\") pod \"metallb-operator-webhook-server-55494b4d7-h6xch\" (UID: \"f83986c8-b147-48b9-84ec-f09d01ddd8ce\") " pod="metallb-system/metallb-operator-webhook-server-55494b4d7-h6xch" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.560347 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-55494b4d7-h6xch" Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.721566 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-56666fbc9b-dj55n"] Mar 17 09:25:43 crc kubenswrapper[4813]: W0317 09:25:43.743463 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda63b7bd8_1a98_40d9_9f56_c7662d229b3d.slice/crio-4258506c547b0629d09313ccaa627b23dc853b6aa9657b145f939a30005190c6 WatchSource:0}: Error finding container 4258506c547b0629d09313ccaa627b23dc853b6aa9657b145f939a30005190c6: Status 404 returned error can't find the container with id 4258506c547b0629d09313ccaa627b23dc853b6aa9657b145f939a30005190c6 Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.780527 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-55494b4d7-h6xch"] Mar 17 09:25:43 crc kubenswrapper[4813]: W0317 09:25:43.789663 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf83986c8_b147_48b9_84ec_f09d01ddd8ce.slice/crio-3df026903118d982893903ba6269f18fd771b21826a9db28a7257ce16768b707 WatchSource:0}: Error finding container 3df026903118d982893903ba6269f18fd771b21826a9db28a7257ce16768b707: Status 404 returned error can't find the container with id 3df026903118d982893903ba6269f18fd771b21826a9db28a7257ce16768b707 Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.856333 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-56666fbc9b-dj55n" event={"ID":"a63b7bd8-1a98-40d9-9f56-c7662d229b3d","Type":"ContainerStarted","Data":"4258506c547b0629d09313ccaa627b23dc853b6aa9657b145f939a30005190c6"} Mar 17 09:25:43 crc kubenswrapper[4813]: I0317 09:25:43.857727 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-55494b4d7-h6xch" event={"ID":"f83986c8-b147-48b9-84ec-f09d01ddd8ce","Type":"ContainerStarted","Data":"3df026903118d982893903ba6269f18fd771b21826a9db28a7257ce16768b707"} Mar 17 09:25:48 crc kubenswrapper[4813]: I0317 09:25:48.897528 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-55494b4d7-h6xch" event={"ID":"f83986c8-b147-48b9-84ec-f09d01ddd8ce","Type":"ContainerStarted","Data":"77feb398c62c9d13438690090835445a3a453f528d11d0dfb4e455d4ce6d59dd"} Mar 17 09:25:48 crc kubenswrapper[4813]: I0317 09:25:48.897939 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-55494b4d7-h6xch" Mar 17 09:25:48 crc kubenswrapper[4813]: I0317 09:25:48.899318 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-56666fbc9b-dj55n" event={"ID":"a63b7bd8-1a98-40d9-9f56-c7662d229b3d","Type":"ContainerStarted","Data":"cbde879206daacf7ebe9297dba41af638f5e7181305d8c3945f23efe203b213f"} Mar 17 09:25:48 crc kubenswrapper[4813]: I0317 09:25:48.899468 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-56666fbc9b-dj55n" Mar 17 09:25:48 crc kubenswrapper[4813]: I0317 09:25:48.932496 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-55494b4d7-h6xch" podStartSLOduration=1.63659218 podStartE2EDuration="5.932476983s" podCreationTimestamp="2026-03-17 09:25:43 +0000 UTC" firstStartedPulling="2026-03-17 09:25:43.792495387 +0000 UTC m=+965.893298886" lastFinishedPulling="2026-03-17 09:25:48.08838019 +0000 UTC m=+970.189183689" observedRunningTime="2026-03-17 09:25:48.912767017 +0000 UTC m=+971.013570516" watchObservedRunningTime="2026-03-17 09:25:48.932476983 +0000 UTC m=+971.033280482" Mar 17 09:25:48 crc kubenswrapper[4813]: I0317 09:25:48.934094 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-56666fbc9b-dj55n" podStartSLOduration=2.604480661 podStartE2EDuration="6.934083954s" podCreationTimestamp="2026-03-17 09:25:42 +0000 UTC" firstStartedPulling="2026-03-17 09:25:43.750271817 +0000 UTC m=+965.851075316" lastFinishedPulling="2026-03-17 09:25:48.07987511 +0000 UTC m=+970.180678609" observedRunningTime="2026-03-17 09:25:48.931672238 +0000 UTC m=+971.032475747" watchObservedRunningTime="2026-03-17 09:25:48.934083954 +0000 UTC m=+971.034887453" Mar 17 09:26:00 crc kubenswrapper[4813]: I0317 09:26:00.130986 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562326-rbvpg"] Mar 17 09:26:00 crc kubenswrapper[4813]: I0317 09:26:00.132361 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562326-rbvpg" Mar 17 09:26:00 crc kubenswrapper[4813]: I0317 09:26:00.134433 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:26:00 crc kubenswrapper[4813]: I0317 09:26:00.134435 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:26:00 crc kubenswrapper[4813]: I0317 09:26:00.134784 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:26:00 crc kubenswrapper[4813]: I0317 09:26:00.139358 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562326-rbvpg"] Mar 17 09:26:00 crc kubenswrapper[4813]: I0317 09:26:00.228327 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t2n5m\" (UniqueName: \"kubernetes.io/projected/d5793f57-43fd-4c88-b61a-50702ff099ce-kube-api-access-t2n5m\") pod \"auto-csr-approver-29562326-rbvpg\" (UID: \"d5793f57-43fd-4c88-b61a-50702ff099ce\") " pod="openshift-infra/auto-csr-approver-29562326-rbvpg" Mar 17 09:26:00 crc kubenswrapper[4813]: I0317 09:26:00.329887 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t2n5m\" (UniqueName: \"kubernetes.io/projected/d5793f57-43fd-4c88-b61a-50702ff099ce-kube-api-access-t2n5m\") pod \"auto-csr-approver-29562326-rbvpg\" (UID: \"d5793f57-43fd-4c88-b61a-50702ff099ce\") " pod="openshift-infra/auto-csr-approver-29562326-rbvpg" Mar 17 09:26:00 crc kubenswrapper[4813]: I0317 09:26:00.372527 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t2n5m\" (UniqueName: \"kubernetes.io/projected/d5793f57-43fd-4c88-b61a-50702ff099ce-kube-api-access-t2n5m\") pod \"auto-csr-approver-29562326-rbvpg\" (UID: \"d5793f57-43fd-4c88-b61a-50702ff099ce\") " pod="openshift-infra/auto-csr-approver-29562326-rbvpg" Mar 17 09:26:00 crc kubenswrapper[4813]: I0317 09:26:00.478704 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562326-rbvpg" Mar 17 09:26:00 crc kubenswrapper[4813]: I0317 09:26:00.961758 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562326-rbvpg"] Mar 17 09:26:00 crc kubenswrapper[4813]: I0317 09:26:00.985997 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562326-rbvpg" event={"ID":"d5793f57-43fd-4c88-b61a-50702ff099ce","Type":"ContainerStarted","Data":"bf19dedbf20014b076b9cfe3810104dce10806a3251fe6251b24d2599988bd6b"} Mar 17 09:26:03 crc kubenswrapper[4813]: I0317 09:26:03.002037 4813 generic.go:334] "Generic (PLEG): container finished" podID="d5793f57-43fd-4c88-b61a-50702ff099ce" containerID="60a4c6e72559b0d6e67762dad3d98f179738de32e3fbf9bfcb1db03da4733ced" exitCode=0 Mar 17 09:26:03 crc kubenswrapper[4813]: I0317 09:26:03.002255 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562326-rbvpg" event={"ID":"d5793f57-43fd-4c88-b61a-50702ff099ce","Type":"ContainerDied","Data":"60a4c6e72559b0d6e67762dad3d98f179738de32e3fbf9bfcb1db03da4733ced"} Mar 17 09:26:03 crc kubenswrapper[4813]: I0317 09:26:03.566321 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-55494b4d7-h6xch" Mar 17 09:26:04 crc kubenswrapper[4813]: I0317 09:26:04.273992 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562326-rbvpg" Mar 17 09:26:04 crc kubenswrapper[4813]: I0317 09:26:04.390743 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t2n5m\" (UniqueName: \"kubernetes.io/projected/d5793f57-43fd-4c88-b61a-50702ff099ce-kube-api-access-t2n5m\") pod \"d5793f57-43fd-4c88-b61a-50702ff099ce\" (UID: \"d5793f57-43fd-4c88-b61a-50702ff099ce\") " Mar 17 09:26:04 crc kubenswrapper[4813]: I0317 09:26:04.396737 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d5793f57-43fd-4c88-b61a-50702ff099ce-kube-api-access-t2n5m" (OuterVolumeSpecName: "kube-api-access-t2n5m") pod "d5793f57-43fd-4c88-b61a-50702ff099ce" (UID: "d5793f57-43fd-4c88-b61a-50702ff099ce"). InnerVolumeSpecName "kube-api-access-t2n5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:26:04 crc kubenswrapper[4813]: I0317 09:26:04.492435 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t2n5m\" (UniqueName: \"kubernetes.io/projected/d5793f57-43fd-4c88-b61a-50702ff099ce-kube-api-access-t2n5m\") on node \"crc\" DevicePath \"\"" Mar 17 09:26:05 crc kubenswrapper[4813]: I0317 09:26:05.015763 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562326-rbvpg" event={"ID":"d5793f57-43fd-4c88-b61a-50702ff099ce","Type":"ContainerDied","Data":"bf19dedbf20014b076b9cfe3810104dce10806a3251fe6251b24d2599988bd6b"} Mar 17 09:26:05 crc kubenswrapper[4813]: I0317 09:26:05.015799 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bf19dedbf20014b076b9cfe3810104dce10806a3251fe6251b24d2599988bd6b" Mar 17 09:26:05 crc kubenswrapper[4813]: I0317 09:26:05.015848 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562326-rbvpg" Mar 17 09:26:05 crc kubenswrapper[4813]: I0317 09:26:05.331740 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562320-bfknz"] Mar 17 09:26:05 crc kubenswrapper[4813]: I0317 09:26:05.339300 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562320-bfknz"] Mar 17 09:26:06 crc kubenswrapper[4813]: I0317 09:26:06.743197 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75bf03ea-aeae-4e4b-b774-e4f1b45981d2" path="/var/lib/kubelet/pods/75bf03ea-aeae-4e4b-b774-e4f1b45981d2/volumes" Mar 17 09:26:14 crc kubenswrapper[4813]: I0317 09:26:14.114850 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:26:14 crc kubenswrapper[4813]: I0317 09:26:14.115649 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.200974 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-56666fbc9b-dj55n" Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.853271 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-wjgfj"] Mar 17 09:26:23 crc kubenswrapper[4813]: E0317 09:26:23.854111 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d5793f57-43fd-4c88-b61a-50702ff099ce" containerName="oc" Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.854138 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d5793f57-43fd-4c88-b61a-50702ff099ce" containerName="oc" Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.854357 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d5793f57-43fd-4c88-b61a-50702ff099ce" containerName="oc" Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.855045 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-wjgfj" Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.857914 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.858191 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-8fl29" Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.858331 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-hj2sl"] Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.860989 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.863126 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.863191 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.869205 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-wjgfj"] Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.949150 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-hzk8t"] Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.949976 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hzk8t" Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.951942 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.952689 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.952817 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-hvjm5" Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.952868 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.976470 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-7bb4cc7c98-5cxrm"] Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.977655 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-5cxrm" Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.979140 4813 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Mar 17 09:26:23 crc kubenswrapper[4813]: I0317 09:26:23.989362 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-5cxrm"] Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.053352 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fl4tb\" (UniqueName: \"kubernetes.io/projected/fbcdac92-95fa-4d87-bc00-fccd03f4c660-kube-api-access-fl4tb\") pod \"frr-k8s-hj2sl\" (UID: \"fbcdac92-95fa-4d87-bc00-fccd03f4c660\") " pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.053408 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ece7eee8-f37e-43f4-a762-986d7d1ae876-metallb-excludel2\") pod \"speaker-hzk8t\" (UID: \"ece7eee8-f37e-43f4-a762-986d7d1ae876\") " pod="metallb-system/speaker-hzk8t" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.053477 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/30c2f61b-c33a-49c4-bde0-ffe8593f29b8-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-wjgfj\" (UID: \"30c2f61b-c33a-49c4-bde0-ffe8593f29b8\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-wjgfj" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.053567 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ece7eee8-f37e-43f4-a762-986d7d1ae876-metrics-certs\") pod \"speaker-hzk8t\" (UID: \"ece7eee8-f37e-43f4-a762-986d7d1ae876\") " pod="metallb-system/speaker-hzk8t" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.053671 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/fbcdac92-95fa-4d87-bc00-fccd03f4c660-frr-sockets\") pod \"frr-k8s-hj2sl\" (UID: \"fbcdac92-95fa-4d87-bc00-fccd03f4c660\") " pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.053693 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/fbcdac92-95fa-4d87-bc00-fccd03f4c660-reloader\") pod \"frr-k8s-hj2sl\" (UID: \"fbcdac92-95fa-4d87-bc00-fccd03f4c660\") " pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.053722 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/fbcdac92-95fa-4d87-bc00-fccd03f4c660-frr-conf\") pod \"frr-k8s-hj2sl\" (UID: \"fbcdac92-95fa-4d87-bc00-fccd03f4c660\") " pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.053747 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ece7eee8-f37e-43f4-a762-986d7d1ae876-memberlist\") pod \"speaker-hzk8t\" (UID: \"ece7eee8-f37e-43f4-a762-986d7d1ae876\") " pod="metallb-system/speaker-hzk8t" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.053858 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnx5j\" (UniqueName: \"kubernetes.io/projected/ece7eee8-f37e-43f4-a762-986d7d1ae876-kube-api-access-xnx5j\") pod \"speaker-hzk8t\" (UID: \"ece7eee8-f37e-43f4-a762-986d7d1ae876\") " pod="metallb-system/speaker-hzk8t" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.053880 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fbcdac92-95fa-4d87-bc00-fccd03f4c660-metrics-certs\") pod \"frr-k8s-hj2sl\" (UID: \"fbcdac92-95fa-4d87-bc00-fccd03f4c660\") " pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.053944 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/fbcdac92-95fa-4d87-bc00-fccd03f4c660-metrics\") pod \"frr-k8s-hj2sl\" (UID: \"fbcdac92-95fa-4d87-bc00-fccd03f4c660\") " pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.054238 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/fbcdac92-95fa-4d87-bc00-fccd03f4c660-frr-startup\") pod \"frr-k8s-hj2sl\" (UID: \"fbcdac92-95fa-4d87-bc00-fccd03f4c660\") " pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.054415 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lg6tn\" (UniqueName: \"kubernetes.io/projected/30c2f61b-c33a-49c4-bde0-ffe8593f29b8-kube-api-access-lg6tn\") pod \"frr-k8s-webhook-server-bcc4b6f68-wjgfj\" (UID: \"30c2f61b-c33a-49c4-bde0-ffe8593f29b8\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-wjgfj" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.155320 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ece7eee8-f37e-43f4-a762-986d7d1ae876-memberlist\") pod \"speaker-hzk8t\" (UID: \"ece7eee8-f37e-43f4-a762-986d7d1ae876\") " pod="metallb-system/speaker-hzk8t" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.155383 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnx5j\" (UniqueName: \"kubernetes.io/projected/ece7eee8-f37e-43f4-a762-986d7d1ae876-kube-api-access-xnx5j\") pod \"speaker-hzk8t\" (UID: \"ece7eee8-f37e-43f4-a762-986d7d1ae876\") " pod="metallb-system/speaker-hzk8t" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.155407 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fbcdac92-95fa-4d87-bc00-fccd03f4c660-metrics-certs\") pod \"frr-k8s-hj2sl\" (UID: \"fbcdac92-95fa-4d87-bc00-fccd03f4c660\") " pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.155448 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/fbcdac92-95fa-4d87-bc00-fccd03f4c660-metrics\") pod \"frr-k8s-hj2sl\" (UID: \"fbcdac92-95fa-4d87-bc00-fccd03f4c660\") " pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.155475 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/fbcdac92-95fa-4d87-bc00-fccd03f4c660-frr-startup\") pod \"frr-k8s-hj2sl\" (UID: \"fbcdac92-95fa-4d87-bc00-fccd03f4c660\") " pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.155497 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lg6tn\" (UniqueName: \"kubernetes.io/projected/30c2f61b-c33a-49c4-bde0-ffe8593f29b8-kube-api-access-lg6tn\") pod \"frr-k8s-webhook-server-bcc4b6f68-wjgfj\" (UID: \"30c2f61b-c33a-49c4-bde0-ffe8593f29b8\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-wjgfj" Mar 17 09:26:24 crc kubenswrapper[4813]: E0317 09:26:24.155510 4813 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 17 09:26:24 crc kubenswrapper[4813]: E0317 09:26:24.155577 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ece7eee8-f37e-43f4-a762-986d7d1ae876-memberlist podName:ece7eee8-f37e-43f4-a762-986d7d1ae876 nodeName:}" failed. No retries permitted until 2026-03-17 09:26:24.655556883 +0000 UTC m=+1006.756360382 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/ece7eee8-f37e-43f4-a762-986d7d1ae876-memberlist") pod "speaker-hzk8t" (UID: "ece7eee8-f37e-43f4-a762-986d7d1ae876") : secret "metallb-memberlist" not found Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.155521 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fl4tb\" (UniqueName: \"kubernetes.io/projected/fbcdac92-95fa-4d87-bc00-fccd03f4c660-kube-api-access-fl4tb\") pod \"frr-k8s-hj2sl\" (UID: \"fbcdac92-95fa-4d87-bc00-fccd03f4c660\") " pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.155762 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ece7eee8-f37e-43f4-a762-986d7d1ae876-metallb-excludel2\") pod \"speaker-hzk8t\" (UID: \"ece7eee8-f37e-43f4-a762-986d7d1ae876\") " pod="metallb-system/speaker-hzk8t" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.155829 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8503b49f-ae8d-4e0d-b3c6-30320e0e283a-metrics-certs\") pod \"controller-7bb4cc7c98-5cxrm\" (UID: \"8503b49f-ae8d-4e0d-b3c6-30320e0e283a\") " pod="metallb-system/controller-7bb4cc7c98-5cxrm" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.155853 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxsps\" (UniqueName: \"kubernetes.io/projected/8503b49f-ae8d-4e0d-b3c6-30320e0e283a-kube-api-access-xxsps\") pod \"controller-7bb4cc7c98-5cxrm\" (UID: \"8503b49f-ae8d-4e0d-b3c6-30320e0e283a\") " pod="metallb-system/controller-7bb4cc7c98-5cxrm" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.155913 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/30c2f61b-c33a-49c4-bde0-ffe8593f29b8-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-wjgfj\" (UID: \"30c2f61b-c33a-49c4-bde0-ffe8593f29b8\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-wjgfj" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.155917 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/fbcdac92-95fa-4d87-bc00-fccd03f4c660-metrics\") pod \"frr-k8s-hj2sl\" (UID: \"fbcdac92-95fa-4d87-bc00-fccd03f4c660\") " pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.155938 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8503b49f-ae8d-4e0d-b3c6-30320e0e283a-cert\") pod \"controller-7bb4cc7c98-5cxrm\" (UID: \"8503b49f-ae8d-4e0d-b3c6-30320e0e283a\") " pod="metallb-system/controller-7bb4cc7c98-5cxrm" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.156054 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ece7eee8-f37e-43f4-a762-986d7d1ae876-metrics-certs\") pod \"speaker-hzk8t\" (UID: \"ece7eee8-f37e-43f4-a762-986d7d1ae876\") " pod="metallb-system/speaker-hzk8t" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.156088 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/fbcdac92-95fa-4d87-bc00-fccd03f4c660-frr-sockets\") pod \"frr-k8s-hj2sl\" (UID: \"fbcdac92-95fa-4d87-bc00-fccd03f4c660\") " pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.156108 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/fbcdac92-95fa-4d87-bc00-fccd03f4c660-reloader\") pod \"frr-k8s-hj2sl\" (UID: \"fbcdac92-95fa-4d87-bc00-fccd03f4c660\") " pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.156128 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/fbcdac92-95fa-4d87-bc00-fccd03f4c660-frr-conf\") pod \"frr-k8s-hj2sl\" (UID: \"fbcdac92-95fa-4d87-bc00-fccd03f4c660\") " pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.156412 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/fbcdac92-95fa-4d87-bc00-fccd03f4c660-frr-conf\") pod \"frr-k8s-hj2sl\" (UID: \"fbcdac92-95fa-4d87-bc00-fccd03f4c660\") " pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.156642 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/fbcdac92-95fa-4d87-bc00-fccd03f4c660-frr-sockets\") pod \"frr-k8s-hj2sl\" (UID: \"fbcdac92-95fa-4d87-bc00-fccd03f4c660\") " pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.156777 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/ece7eee8-f37e-43f4-a762-986d7d1ae876-metallb-excludel2\") pod \"speaker-hzk8t\" (UID: \"ece7eee8-f37e-43f4-a762-986d7d1ae876\") " pod="metallb-system/speaker-hzk8t" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.156780 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/fbcdac92-95fa-4d87-bc00-fccd03f4c660-frr-startup\") pod \"frr-k8s-hj2sl\" (UID: \"fbcdac92-95fa-4d87-bc00-fccd03f4c660\") " pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.156819 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/fbcdac92-95fa-4d87-bc00-fccd03f4c660-reloader\") pod \"frr-k8s-hj2sl\" (UID: \"fbcdac92-95fa-4d87-bc00-fccd03f4c660\") " pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.163130 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/fbcdac92-95fa-4d87-bc00-fccd03f4c660-metrics-certs\") pod \"frr-k8s-hj2sl\" (UID: \"fbcdac92-95fa-4d87-bc00-fccd03f4c660\") " pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.163520 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/30c2f61b-c33a-49c4-bde0-ffe8593f29b8-cert\") pod \"frr-k8s-webhook-server-bcc4b6f68-wjgfj\" (UID: \"30c2f61b-c33a-49c4-bde0-ffe8593f29b8\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-wjgfj" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.170533 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ece7eee8-f37e-43f4-a762-986d7d1ae876-metrics-certs\") pod \"speaker-hzk8t\" (UID: \"ece7eee8-f37e-43f4-a762-986d7d1ae876\") " pod="metallb-system/speaker-hzk8t" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.176146 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnx5j\" (UniqueName: \"kubernetes.io/projected/ece7eee8-f37e-43f4-a762-986d7d1ae876-kube-api-access-xnx5j\") pod \"speaker-hzk8t\" (UID: \"ece7eee8-f37e-43f4-a762-986d7d1ae876\") " pod="metallb-system/speaker-hzk8t" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.177442 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lg6tn\" (UniqueName: \"kubernetes.io/projected/30c2f61b-c33a-49c4-bde0-ffe8593f29b8-kube-api-access-lg6tn\") pod \"frr-k8s-webhook-server-bcc4b6f68-wjgfj\" (UID: \"30c2f61b-c33a-49c4-bde0-ffe8593f29b8\") " pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-wjgfj" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.181170 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fl4tb\" (UniqueName: \"kubernetes.io/projected/fbcdac92-95fa-4d87-bc00-fccd03f4c660-kube-api-access-fl4tb\") pod \"frr-k8s-hj2sl\" (UID: \"fbcdac92-95fa-4d87-bc00-fccd03f4c660\") " pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.185956 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.256486 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8503b49f-ae8d-4e0d-b3c6-30320e0e283a-cert\") pod \"controller-7bb4cc7c98-5cxrm\" (UID: \"8503b49f-ae8d-4e0d-b3c6-30320e0e283a\") " pod="metallb-system/controller-7bb4cc7c98-5cxrm" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.256666 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8503b49f-ae8d-4e0d-b3c6-30320e0e283a-metrics-certs\") pod \"controller-7bb4cc7c98-5cxrm\" (UID: \"8503b49f-ae8d-4e0d-b3c6-30320e0e283a\") " pod="metallb-system/controller-7bb4cc7c98-5cxrm" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.256705 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxsps\" (UniqueName: \"kubernetes.io/projected/8503b49f-ae8d-4e0d-b3c6-30320e0e283a-kube-api-access-xxsps\") pod \"controller-7bb4cc7c98-5cxrm\" (UID: \"8503b49f-ae8d-4e0d-b3c6-30320e0e283a\") " pod="metallb-system/controller-7bb4cc7c98-5cxrm" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.259752 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/8503b49f-ae8d-4e0d-b3c6-30320e0e283a-metrics-certs\") pod \"controller-7bb4cc7c98-5cxrm\" (UID: \"8503b49f-ae8d-4e0d-b3c6-30320e0e283a\") " pod="metallb-system/controller-7bb4cc7c98-5cxrm" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.264079 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8503b49f-ae8d-4e0d-b3c6-30320e0e283a-cert\") pod \"controller-7bb4cc7c98-5cxrm\" (UID: \"8503b49f-ae8d-4e0d-b3c6-30320e0e283a\") " pod="metallb-system/controller-7bb4cc7c98-5cxrm" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.278957 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxsps\" (UniqueName: \"kubernetes.io/projected/8503b49f-ae8d-4e0d-b3c6-30320e0e283a-kube-api-access-xxsps\") pod \"controller-7bb4cc7c98-5cxrm\" (UID: \"8503b49f-ae8d-4e0d-b3c6-30320e0e283a\") " pod="metallb-system/controller-7bb4cc7c98-5cxrm" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.291626 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-7bb4cc7c98-5cxrm" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.476361 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-wjgfj" Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.677670 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ece7eee8-f37e-43f4-a762-986d7d1ae876-memberlist\") pod \"speaker-hzk8t\" (UID: \"ece7eee8-f37e-43f4-a762-986d7d1ae876\") " pod="metallb-system/speaker-hzk8t" Mar 17 09:26:24 crc kubenswrapper[4813]: E0317 09:26:24.677833 4813 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Mar 17 09:26:24 crc kubenswrapper[4813]: E0317 09:26:24.678181 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ece7eee8-f37e-43f4-a762-986d7d1ae876-memberlist podName:ece7eee8-f37e-43f4-a762-986d7d1ae876 nodeName:}" failed. No retries permitted until 2026-03-17 09:26:25.678154691 +0000 UTC m=+1007.778958190 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/ece7eee8-f37e-43f4-a762-986d7d1ae876-memberlist") pod "speaker-hzk8t" (UID: "ece7eee8-f37e-43f4-a762-986d7d1ae876") : secret "metallb-memberlist" not found Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.776302 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-7bb4cc7c98-5cxrm"] Mar 17 09:26:24 crc kubenswrapper[4813]: W0317 09:26:24.781216 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8503b49f_ae8d_4e0d_b3c6_30320e0e283a.slice/crio-ca0678fa04179729a0565f5367c4315acfff61f73cbb6d03c8edeb9fb03ae51c WatchSource:0}: Error finding container ca0678fa04179729a0565f5367c4315acfff61f73cbb6d03c8edeb9fb03ae51c: Status 404 returned error can't find the container with id ca0678fa04179729a0565f5367c4315acfff61f73cbb6d03c8edeb9fb03ae51c Mar 17 09:26:24 crc kubenswrapper[4813]: I0317 09:26:24.881680 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-bcc4b6f68-wjgfj"] Mar 17 09:26:24 crc kubenswrapper[4813]: W0317 09:26:24.887752 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30c2f61b_c33a_49c4_bde0_ffe8593f29b8.slice/crio-f22bd62b80416ac0171e35e467099310162a50faa080c6ae6e352360306e3a4f WatchSource:0}: Error finding container f22bd62b80416ac0171e35e467099310162a50faa080c6ae6e352360306e3a4f: Status 404 returned error can't find the container with id f22bd62b80416ac0171e35e467099310162a50faa080c6ae6e352360306e3a4f Mar 17 09:26:25 crc kubenswrapper[4813]: I0317 09:26:25.153087 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hj2sl" event={"ID":"fbcdac92-95fa-4d87-bc00-fccd03f4c660","Type":"ContainerStarted","Data":"4436f392f9f19cd5b123977818dad9d1ffa29e1abcd7d05a78ef145681fb0036"} Mar 17 09:26:25 crc kubenswrapper[4813]: I0317 09:26:25.154304 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-wjgfj" event={"ID":"30c2f61b-c33a-49c4-bde0-ffe8593f29b8","Type":"ContainerStarted","Data":"f22bd62b80416ac0171e35e467099310162a50faa080c6ae6e352360306e3a4f"} Mar 17 09:26:25 crc kubenswrapper[4813]: I0317 09:26:25.156088 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-5cxrm" event={"ID":"8503b49f-ae8d-4e0d-b3c6-30320e0e283a","Type":"ContainerStarted","Data":"6a97ffa960f23beeff640e127a9f476a460ee99c0184ef8dd7c2e85e2c575e8a"} Mar 17 09:26:25 crc kubenswrapper[4813]: I0317 09:26:25.156118 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-5cxrm" event={"ID":"8503b49f-ae8d-4e0d-b3c6-30320e0e283a","Type":"ContainerStarted","Data":"7a5121c5e9eb461267e6f2c4513b51feff26ffbe7e6f769733b106241f9b8b56"} Mar 17 09:26:25 crc kubenswrapper[4813]: I0317 09:26:25.156133 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-7bb4cc7c98-5cxrm" event={"ID":"8503b49f-ae8d-4e0d-b3c6-30320e0e283a","Type":"ContainerStarted","Data":"ca0678fa04179729a0565f5367c4315acfff61f73cbb6d03c8edeb9fb03ae51c"} Mar 17 09:26:25 crc kubenswrapper[4813]: I0317 09:26:25.157091 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-7bb4cc7c98-5cxrm" Mar 17 09:26:25 crc kubenswrapper[4813]: I0317 09:26:25.177289 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-7bb4cc7c98-5cxrm" podStartSLOduration=2.177263424 podStartE2EDuration="2.177263424s" podCreationTimestamp="2026-03-17 09:26:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:26:25.172435841 +0000 UTC m=+1007.273239380" watchObservedRunningTime="2026-03-17 09:26:25.177263424 +0000 UTC m=+1007.278066943" Mar 17 09:26:25 crc kubenswrapper[4813]: I0317 09:26:25.691142 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ece7eee8-f37e-43f4-a762-986d7d1ae876-memberlist\") pod \"speaker-hzk8t\" (UID: \"ece7eee8-f37e-43f4-a762-986d7d1ae876\") " pod="metallb-system/speaker-hzk8t" Mar 17 09:26:25 crc kubenswrapper[4813]: I0317 09:26:25.696961 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/ece7eee8-f37e-43f4-a762-986d7d1ae876-memberlist\") pod \"speaker-hzk8t\" (UID: \"ece7eee8-f37e-43f4-a762-986d7d1ae876\") " pod="metallb-system/speaker-hzk8t" Mar 17 09:26:25 crc kubenswrapper[4813]: I0317 09:26:25.765010 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-hzk8t" Mar 17 09:26:26 crc kubenswrapper[4813]: I0317 09:26:26.167579 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hzk8t" event={"ID":"ece7eee8-f37e-43f4-a762-986d7d1ae876","Type":"ContainerStarted","Data":"5b6d74222762d0c765c43bd1855818c39a64772984a5669811ae2001bae271c4"} Mar 17 09:26:26 crc kubenswrapper[4813]: I0317 09:26:26.168014 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hzk8t" event={"ID":"ece7eee8-f37e-43f4-a762-986d7d1ae876","Type":"ContainerStarted","Data":"46fe9ed2dd88b22509f48534ac05936a85187e72319c1c4471df0004667af643"} Mar 17 09:26:27 crc kubenswrapper[4813]: I0317 09:26:27.178199 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-hzk8t" event={"ID":"ece7eee8-f37e-43f4-a762-986d7d1ae876","Type":"ContainerStarted","Data":"0904b8b3ed69a810f568bb4302b2a9efe4fcb95029cc4a44e0838b1095e6c7ee"} Mar 17 09:26:27 crc kubenswrapper[4813]: I0317 09:26:27.178261 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-hzk8t" Mar 17 09:26:27 crc kubenswrapper[4813]: I0317 09:26:27.201935 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-hzk8t" podStartSLOduration=4.201915082 podStartE2EDuration="4.201915082s" podCreationTimestamp="2026-03-17 09:26:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:26:27.195487637 +0000 UTC m=+1009.296291146" watchObservedRunningTime="2026-03-17 09:26:27.201915082 +0000 UTC m=+1009.302718581" Mar 17 09:26:32 crc kubenswrapper[4813]: I0317 09:26:32.217510 4813 generic.go:334] "Generic (PLEG): container finished" podID="fbcdac92-95fa-4d87-bc00-fccd03f4c660" containerID="fb57e0da25a0ff9930c7b371eb1e930bcd66e8eade33884bcc6d034a3563feba" exitCode=0 Mar 17 09:26:32 crc kubenswrapper[4813]: I0317 09:26:32.217568 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hj2sl" event={"ID":"fbcdac92-95fa-4d87-bc00-fccd03f4c660","Type":"ContainerDied","Data":"fb57e0da25a0ff9930c7b371eb1e930bcd66e8eade33884bcc6d034a3563feba"} Mar 17 09:26:32 crc kubenswrapper[4813]: I0317 09:26:32.220644 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-wjgfj" event={"ID":"30c2f61b-c33a-49c4-bde0-ffe8593f29b8","Type":"ContainerStarted","Data":"cd2807410bb02c0379d11c180f1b3652ca7d6a110d612c475d0402f4dc8a011c"} Mar 17 09:26:32 crc kubenswrapper[4813]: I0317 09:26:32.220785 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-wjgfj" Mar 17 09:26:32 crc kubenswrapper[4813]: I0317 09:26:32.304242 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-wjgfj" podStartSLOduration=3.1055635600000002 podStartE2EDuration="9.304216481s" podCreationTimestamp="2026-03-17 09:26:23 +0000 UTC" firstStartedPulling="2026-03-17 09:26:24.889996465 +0000 UTC m=+1006.990799964" lastFinishedPulling="2026-03-17 09:26:31.088649366 +0000 UTC m=+1013.189452885" observedRunningTime="2026-03-17 09:26:32.301254317 +0000 UTC m=+1014.402057826" watchObservedRunningTime="2026-03-17 09:26:32.304216481 +0000 UTC m=+1014.405019990" Mar 17 09:26:33 crc kubenswrapper[4813]: I0317 09:26:33.231381 4813 generic.go:334] "Generic (PLEG): container finished" podID="fbcdac92-95fa-4d87-bc00-fccd03f4c660" containerID="4b76ef7423866103f6d8d053038583354ef6f2cd6e05dc0a53c0d1fd68af9e10" exitCode=0 Mar 17 09:26:33 crc kubenswrapper[4813]: I0317 09:26:33.231478 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hj2sl" event={"ID":"fbcdac92-95fa-4d87-bc00-fccd03f4c660","Type":"ContainerDied","Data":"4b76ef7423866103f6d8d053038583354ef6f2cd6e05dc0a53c0d1fd68af9e10"} Mar 17 09:26:34 crc kubenswrapper[4813]: I0317 09:26:34.256296 4813 generic.go:334] "Generic (PLEG): container finished" podID="fbcdac92-95fa-4d87-bc00-fccd03f4c660" containerID="19a5c63177444cc9737fc62b7c75843f8947724a6689d2be8cfe04b42bd1db4d" exitCode=0 Mar 17 09:26:34 crc kubenswrapper[4813]: I0317 09:26:34.256359 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hj2sl" event={"ID":"fbcdac92-95fa-4d87-bc00-fccd03f4c660","Type":"ContainerDied","Data":"19a5c63177444cc9737fc62b7c75843f8947724a6689d2be8cfe04b42bd1db4d"} Mar 17 09:26:34 crc kubenswrapper[4813]: I0317 09:26:34.296106 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-7bb4cc7c98-5cxrm" Mar 17 09:26:35 crc kubenswrapper[4813]: I0317 09:26:35.273557 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hj2sl" event={"ID":"fbcdac92-95fa-4d87-bc00-fccd03f4c660","Type":"ContainerStarted","Data":"5f7d835aeb04b72c1590fec68d1df0308fd114d5b80043295161db5203c89329"} Mar 17 09:26:35 crc kubenswrapper[4813]: I0317 09:26:35.274101 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hj2sl" event={"ID":"fbcdac92-95fa-4d87-bc00-fccd03f4c660","Type":"ContainerStarted","Data":"e162591530d0b84eac30677f6d8aac87bf6e2013de63f7416e564f56f58791a5"} Mar 17 09:26:35 crc kubenswrapper[4813]: I0317 09:26:35.274129 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hj2sl" event={"ID":"fbcdac92-95fa-4d87-bc00-fccd03f4c660","Type":"ContainerStarted","Data":"f79ef5e3e6d058cd246063fb3234e76d461c42a202d90867bf15423ac2399e6e"} Mar 17 09:26:35 crc kubenswrapper[4813]: I0317 09:26:35.274151 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hj2sl" event={"ID":"fbcdac92-95fa-4d87-bc00-fccd03f4c660","Type":"ContainerStarted","Data":"c051d0847ca298f3cdac1f4e98a1b631f256a747a503041c4863be11fa57c412"} Mar 17 09:26:35 crc kubenswrapper[4813]: I0317 09:26:35.274169 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hj2sl" event={"ID":"fbcdac92-95fa-4d87-bc00-fccd03f4c660","Type":"ContainerStarted","Data":"6414338bdea90504f7f3a6545f253ad62dcceb82c1dc0a7549333078658f4b4e"} Mar 17 09:26:36 crc kubenswrapper[4813]: I0317 09:26:36.283671 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-hj2sl" event={"ID":"fbcdac92-95fa-4d87-bc00-fccd03f4c660","Type":"ContainerStarted","Data":"cb7bd33d171c08fb03932bbc45ccac48e603dd0aa45b40deb0465e9348381403"} Mar 17 09:26:36 crc kubenswrapper[4813]: I0317 09:26:36.284648 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:36 crc kubenswrapper[4813]: I0317 09:26:36.319035 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-hj2sl" podStartSLOduration=6.518373991 podStartE2EDuration="13.319014441s" podCreationTimestamp="2026-03-17 09:26:23 +0000 UTC" firstStartedPulling="2026-03-17 09:26:24.317021817 +0000 UTC m=+1006.417825316" lastFinishedPulling="2026-03-17 09:26:31.117662257 +0000 UTC m=+1013.218465766" observedRunningTime="2026-03-17 09:26:36.315679245 +0000 UTC m=+1018.416482754" watchObservedRunningTime="2026-03-17 09:26:36.319014441 +0000 UTC m=+1018.419817940" Mar 17 09:26:39 crc kubenswrapper[4813]: I0317 09:26:39.186193 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:39 crc kubenswrapper[4813]: I0317 09:26:39.235475 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:39 crc kubenswrapper[4813]: I0317 09:26:39.418592 4813 scope.go:117] "RemoveContainer" containerID="14e897629c918331483db943441831f8093d0c3c781e5e96d5e251f70a2a8e54" Mar 17 09:26:44 crc kubenswrapper[4813]: I0317 09:26:44.113902 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:26:44 crc kubenswrapper[4813]: I0317 09:26:44.114310 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:26:44 crc kubenswrapper[4813]: I0317 09:26:44.193034 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-hj2sl" Mar 17 09:26:44 crc kubenswrapper[4813]: I0317 09:26:44.487123 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-bcc4b6f68-wjgfj" Mar 17 09:26:45 crc kubenswrapper[4813]: I0317 09:26:45.004626 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p6mfs"] Mar 17 09:26:45 crc kubenswrapper[4813]: I0317 09:26:45.006159 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p6mfs" Mar 17 09:26:45 crc kubenswrapper[4813]: I0317 09:26:45.023431 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p6mfs"] Mar 17 09:26:45 crc kubenswrapper[4813]: I0317 09:26:45.181899 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/733df085-2100-4070-b18e-46ef5390a164-utilities\") pod \"certified-operators-p6mfs\" (UID: \"733df085-2100-4070-b18e-46ef5390a164\") " pod="openshift-marketplace/certified-operators-p6mfs" Mar 17 09:26:45 crc kubenswrapper[4813]: I0317 09:26:45.181981 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/733df085-2100-4070-b18e-46ef5390a164-catalog-content\") pod \"certified-operators-p6mfs\" (UID: \"733df085-2100-4070-b18e-46ef5390a164\") " pod="openshift-marketplace/certified-operators-p6mfs" Mar 17 09:26:45 crc kubenswrapper[4813]: I0317 09:26:45.182018 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8kn7\" (UniqueName: \"kubernetes.io/projected/733df085-2100-4070-b18e-46ef5390a164-kube-api-access-t8kn7\") pod \"certified-operators-p6mfs\" (UID: \"733df085-2100-4070-b18e-46ef5390a164\") " pod="openshift-marketplace/certified-operators-p6mfs" Mar 17 09:26:45 crc kubenswrapper[4813]: I0317 09:26:45.285940 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/733df085-2100-4070-b18e-46ef5390a164-catalog-content\") pod \"certified-operators-p6mfs\" (UID: \"733df085-2100-4070-b18e-46ef5390a164\") " pod="openshift-marketplace/certified-operators-p6mfs" Mar 17 09:26:45 crc kubenswrapper[4813]: I0317 09:26:45.286015 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8kn7\" (UniqueName: \"kubernetes.io/projected/733df085-2100-4070-b18e-46ef5390a164-kube-api-access-t8kn7\") pod \"certified-operators-p6mfs\" (UID: \"733df085-2100-4070-b18e-46ef5390a164\") " pod="openshift-marketplace/certified-operators-p6mfs" Mar 17 09:26:45 crc kubenswrapper[4813]: I0317 09:26:45.286111 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/733df085-2100-4070-b18e-46ef5390a164-utilities\") pod \"certified-operators-p6mfs\" (UID: \"733df085-2100-4070-b18e-46ef5390a164\") " pod="openshift-marketplace/certified-operators-p6mfs" Mar 17 09:26:45 crc kubenswrapper[4813]: I0317 09:26:45.286566 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/733df085-2100-4070-b18e-46ef5390a164-catalog-content\") pod \"certified-operators-p6mfs\" (UID: \"733df085-2100-4070-b18e-46ef5390a164\") " pod="openshift-marketplace/certified-operators-p6mfs" Mar 17 09:26:45 crc kubenswrapper[4813]: I0317 09:26:45.304346 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/733df085-2100-4070-b18e-46ef5390a164-utilities\") pod \"certified-operators-p6mfs\" (UID: \"733df085-2100-4070-b18e-46ef5390a164\") " pod="openshift-marketplace/certified-operators-p6mfs" Mar 17 09:26:45 crc kubenswrapper[4813]: I0317 09:26:45.339834 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8kn7\" (UniqueName: \"kubernetes.io/projected/733df085-2100-4070-b18e-46ef5390a164-kube-api-access-t8kn7\") pod \"certified-operators-p6mfs\" (UID: \"733df085-2100-4070-b18e-46ef5390a164\") " pod="openshift-marketplace/certified-operators-p6mfs" Mar 17 09:26:45 crc kubenswrapper[4813]: I0317 09:26:45.340334 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p6mfs" Mar 17 09:26:45 crc kubenswrapper[4813]: I0317 09:26:45.769323 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-hzk8t" Mar 17 09:26:45 crc kubenswrapper[4813]: I0317 09:26:45.805168 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p6mfs"] Mar 17 09:26:46 crc kubenswrapper[4813]: I0317 09:26:46.384873 4813 generic.go:334] "Generic (PLEG): container finished" podID="733df085-2100-4070-b18e-46ef5390a164" containerID="723058636df9ee0769fe5a98ac4d950e6f0dcf998a8e7dc007e1df8d1553225d" exitCode=0 Mar 17 09:26:46 crc kubenswrapper[4813]: I0317 09:26:46.384942 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p6mfs" event={"ID":"733df085-2100-4070-b18e-46ef5390a164","Type":"ContainerDied","Data":"723058636df9ee0769fe5a98ac4d950e6f0dcf998a8e7dc007e1df8d1553225d"} Mar 17 09:26:46 crc kubenswrapper[4813]: I0317 09:26:46.385189 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p6mfs" event={"ID":"733df085-2100-4070-b18e-46ef5390a164","Type":"ContainerStarted","Data":"29402866e2a73195d3d6d1cf5cd746ac2d1bc5b5154555022ba6df9ef4c5bb60"} Mar 17 09:26:48 crc kubenswrapper[4813]: I0317 09:26:48.402179 4813 generic.go:334] "Generic (PLEG): container finished" podID="733df085-2100-4070-b18e-46ef5390a164" containerID="c5e25649b3f49002dd6bfbec6c9dd4402d4f085aef758bf4dc449528b2542053" exitCode=0 Mar 17 09:26:48 crc kubenswrapper[4813]: I0317 09:26:48.402251 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p6mfs" event={"ID":"733df085-2100-4070-b18e-46ef5390a164","Type":"ContainerDied","Data":"c5e25649b3f49002dd6bfbec6c9dd4402d4f085aef758bf4dc449528b2542053"} Mar 17 09:26:49 crc kubenswrapper[4813]: I0317 09:26:49.410497 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p6mfs" event={"ID":"733df085-2100-4070-b18e-46ef5390a164","Type":"ContainerStarted","Data":"59e5fa80d9904afc9f18b06d0f3367fb97417b85a7b7c929979a7b3f5dc89cd1"} Mar 17 09:26:49 crc kubenswrapper[4813]: I0317 09:26:49.433403 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p6mfs" podStartSLOduration=2.97120783 podStartE2EDuration="5.433387066s" podCreationTimestamp="2026-03-17 09:26:44 +0000 UTC" firstStartedPulling="2026-03-17 09:26:46.387024986 +0000 UTC m=+1028.487828515" lastFinishedPulling="2026-03-17 09:26:48.849204242 +0000 UTC m=+1030.950007751" observedRunningTime="2026-03-17 09:26:49.429031117 +0000 UTC m=+1031.529834616" watchObservedRunningTime="2026-03-17 09:26:49.433387066 +0000 UTC m=+1031.534190575" Mar 17 09:26:51 crc kubenswrapper[4813]: I0317 09:26:51.996633 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-z8wcr"] Mar 17 09:26:51 crc kubenswrapper[4813]: I0317 09:26:51.997925 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-z8wcr" Mar 17 09:26:52 crc kubenswrapper[4813]: I0317 09:26:52.002286 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Mar 17 09:26:52 crc kubenswrapper[4813]: I0317 09:26:52.002751 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-2b4xz" Mar 17 09:26:52 crc kubenswrapper[4813]: I0317 09:26:52.003189 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Mar 17 09:26:52 crc kubenswrapper[4813]: I0317 09:26:52.025047 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-z8wcr"] Mar 17 09:26:52 crc kubenswrapper[4813]: I0317 09:26:52.101023 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lpvvq\" (UniqueName: \"kubernetes.io/projected/e56922bf-05f4-474c-b883-f1ca16e3e1cd-kube-api-access-lpvvq\") pod \"openstack-operator-index-z8wcr\" (UID: \"e56922bf-05f4-474c-b883-f1ca16e3e1cd\") " pod="openstack-operators/openstack-operator-index-z8wcr" Mar 17 09:26:52 crc kubenswrapper[4813]: I0317 09:26:52.202112 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lpvvq\" (UniqueName: \"kubernetes.io/projected/e56922bf-05f4-474c-b883-f1ca16e3e1cd-kube-api-access-lpvvq\") pod \"openstack-operator-index-z8wcr\" (UID: \"e56922bf-05f4-474c-b883-f1ca16e3e1cd\") " pod="openstack-operators/openstack-operator-index-z8wcr" Mar 17 09:26:52 crc kubenswrapper[4813]: I0317 09:26:52.246062 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lpvvq\" (UniqueName: \"kubernetes.io/projected/e56922bf-05f4-474c-b883-f1ca16e3e1cd-kube-api-access-lpvvq\") pod \"openstack-operator-index-z8wcr\" (UID: \"e56922bf-05f4-474c-b883-f1ca16e3e1cd\") " pod="openstack-operators/openstack-operator-index-z8wcr" Mar 17 09:26:52 crc kubenswrapper[4813]: I0317 09:26:52.331443 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-z8wcr" Mar 17 09:26:52 crc kubenswrapper[4813]: I0317 09:26:52.617701 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-z8wcr"] Mar 17 09:26:53 crc kubenswrapper[4813]: I0317 09:26:53.449926 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-z8wcr" event={"ID":"e56922bf-05f4-474c-b883-f1ca16e3e1cd","Type":"ContainerStarted","Data":"7250b7fd01d2417a9ed429030124e6e70f90035c2aacddcb85f656f4f581ebb4"} Mar 17 09:26:55 crc kubenswrapper[4813]: I0317 09:26:55.340913 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p6mfs" Mar 17 09:26:55 crc kubenswrapper[4813]: I0317 09:26:55.341235 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p6mfs" Mar 17 09:26:55 crc kubenswrapper[4813]: I0317 09:26:55.389497 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p6mfs" Mar 17 09:26:55 crc kubenswrapper[4813]: I0317 09:26:55.532762 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p6mfs" Mar 17 09:26:56 crc kubenswrapper[4813]: I0317 09:26:56.485933 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-z8wcr" event={"ID":"e56922bf-05f4-474c-b883-f1ca16e3e1cd","Type":"ContainerStarted","Data":"c430223a7ae54c30d9983f0de0b8191294b6aa47c6fbd66ab4f1b6321d6deb25"} Mar 17 09:26:56 crc kubenswrapper[4813]: I0317 09:26:56.502002 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-z8wcr" podStartSLOduration=2.564919517 podStartE2EDuration="5.50198779s" podCreationTimestamp="2026-03-17 09:26:51 +0000 UTC" firstStartedPulling="2026-03-17 09:26:52.626067639 +0000 UTC m=+1034.726871138" lastFinishedPulling="2026-03-17 09:26:55.563135912 +0000 UTC m=+1037.663939411" observedRunningTime="2026-03-17 09:26:56.500470471 +0000 UTC m=+1038.601273970" watchObservedRunningTime="2026-03-17 09:26:56.50198779 +0000 UTC m=+1038.602791279" Mar 17 09:26:58 crc kubenswrapper[4813]: I0317 09:26:58.794467 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p6mfs"] Mar 17 09:26:58 crc kubenswrapper[4813]: I0317 09:26:58.794834 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p6mfs" podUID="733df085-2100-4070-b18e-46ef5390a164" containerName="registry-server" containerID="cri-o://59e5fa80d9904afc9f18b06d0f3367fb97417b85a7b7c929979a7b3f5dc89cd1" gracePeriod=2 Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.275670 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p6mfs" Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.412054 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8kn7\" (UniqueName: \"kubernetes.io/projected/733df085-2100-4070-b18e-46ef5390a164-kube-api-access-t8kn7\") pod \"733df085-2100-4070-b18e-46ef5390a164\" (UID: \"733df085-2100-4070-b18e-46ef5390a164\") " Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.412123 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/733df085-2100-4070-b18e-46ef5390a164-catalog-content\") pod \"733df085-2100-4070-b18e-46ef5390a164\" (UID: \"733df085-2100-4070-b18e-46ef5390a164\") " Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.412173 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/733df085-2100-4070-b18e-46ef5390a164-utilities\") pod \"733df085-2100-4070-b18e-46ef5390a164\" (UID: \"733df085-2100-4070-b18e-46ef5390a164\") " Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.413164 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/733df085-2100-4070-b18e-46ef5390a164-utilities" (OuterVolumeSpecName: "utilities") pod "733df085-2100-4070-b18e-46ef5390a164" (UID: "733df085-2100-4070-b18e-46ef5390a164"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.421877 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/733df085-2100-4070-b18e-46ef5390a164-kube-api-access-t8kn7" (OuterVolumeSpecName: "kube-api-access-t8kn7") pod "733df085-2100-4070-b18e-46ef5390a164" (UID: "733df085-2100-4070-b18e-46ef5390a164"). InnerVolumeSpecName "kube-api-access-t8kn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.460102 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/733df085-2100-4070-b18e-46ef5390a164-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "733df085-2100-4070-b18e-46ef5390a164" (UID: "733df085-2100-4070-b18e-46ef5390a164"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.509879 4813 generic.go:334] "Generic (PLEG): container finished" podID="733df085-2100-4070-b18e-46ef5390a164" containerID="59e5fa80d9904afc9f18b06d0f3367fb97417b85a7b7c929979a7b3f5dc89cd1" exitCode=0 Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.509925 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p6mfs" event={"ID":"733df085-2100-4070-b18e-46ef5390a164","Type":"ContainerDied","Data":"59e5fa80d9904afc9f18b06d0f3367fb97417b85a7b7c929979a7b3f5dc89cd1"} Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.509954 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p6mfs" event={"ID":"733df085-2100-4070-b18e-46ef5390a164","Type":"ContainerDied","Data":"29402866e2a73195d3d6d1cf5cd746ac2d1bc5b5154555022ba6df9ef4c5bb60"} Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.509974 4813 scope.go:117] "RemoveContainer" containerID="59e5fa80d9904afc9f18b06d0f3367fb97417b85a7b7c929979a7b3f5dc89cd1" Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.510099 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p6mfs" Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.517158 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8kn7\" (UniqueName: \"kubernetes.io/projected/733df085-2100-4070-b18e-46ef5390a164-kube-api-access-t8kn7\") on node \"crc\" DevicePath \"\"" Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.517187 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/733df085-2100-4070-b18e-46ef5390a164-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.517223 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/733df085-2100-4070-b18e-46ef5390a164-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.534968 4813 scope.go:117] "RemoveContainer" containerID="c5e25649b3f49002dd6bfbec6c9dd4402d4f085aef758bf4dc449528b2542053" Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.545247 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p6mfs"] Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.557220 4813 scope.go:117] "RemoveContainer" containerID="723058636df9ee0769fe5a98ac4d950e6f0dcf998a8e7dc007e1df8d1553225d" Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.559884 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p6mfs"] Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.586902 4813 scope.go:117] "RemoveContainer" containerID="59e5fa80d9904afc9f18b06d0f3367fb97417b85a7b7c929979a7b3f5dc89cd1" Mar 17 09:26:59 crc kubenswrapper[4813]: E0317 09:26:59.587397 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"59e5fa80d9904afc9f18b06d0f3367fb97417b85a7b7c929979a7b3f5dc89cd1\": container with ID starting with 59e5fa80d9904afc9f18b06d0f3367fb97417b85a7b7c929979a7b3f5dc89cd1 not found: ID does not exist" containerID="59e5fa80d9904afc9f18b06d0f3367fb97417b85a7b7c929979a7b3f5dc89cd1" Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.587439 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"59e5fa80d9904afc9f18b06d0f3367fb97417b85a7b7c929979a7b3f5dc89cd1"} err="failed to get container status \"59e5fa80d9904afc9f18b06d0f3367fb97417b85a7b7c929979a7b3f5dc89cd1\": rpc error: code = NotFound desc = could not find container \"59e5fa80d9904afc9f18b06d0f3367fb97417b85a7b7c929979a7b3f5dc89cd1\": container with ID starting with 59e5fa80d9904afc9f18b06d0f3367fb97417b85a7b7c929979a7b3f5dc89cd1 not found: ID does not exist" Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.587465 4813 scope.go:117] "RemoveContainer" containerID="c5e25649b3f49002dd6bfbec6c9dd4402d4f085aef758bf4dc449528b2542053" Mar 17 09:26:59 crc kubenswrapper[4813]: E0317 09:26:59.587872 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5e25649b3f49002dd6bfbec6c9dd4402d4f085aef758bf4dc449528b2542053\": container with ID starting with c5e25649b3f49002dd6bfbec6c9dd4402d4f085aef758bf4dc449528b2542053 not found: ID does not exist" containerID="c5e25649b3f49002dd6bfbec6c9dd4402d4f085aef758bf4dc449528b2542053" Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.587898 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5e25649b3f49002dd6bfbec6c9dd4402d4f085aef758bf4dc449528b2542053"} err="failed to get container status \"c5e25649b3f49002dd6bfbec6c9dd4402d4f085aef758bf4dc449528b2542053\": rpc error: code = NotFound desc = could not find container \"c5e25649b3f49002dd6bfbec6c9dd4402d4f085aef758bf4dc449528b2542053\": container with ID starting with c5e25649b3f49002dd6bfbec6c9dd4402d4f085aef758bf4dc449528b2542053 not found: ID does not exist" Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.587913 4813 scope.go:117] "RemoveContainer" containerID="723058636df9ee0769fe5a98ac4d950e6f0dcf998a8e7dc007e1df8d1553225d" Mar 17 09:26:59 crc kubenswrapper[4813]: E0317 09:26:59.588250 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"723058636df9ee0769fe5a98ac4d950e6f0dcf998a8e7dc007e1df8d1553225d\": container with ID starting with 723058636df9ee0769fe5a98ac4d950e6f0dcf998a8e7dc007e1df8d1553225d not found: ID does not exist" containerID="723058636df9ee0769fe5a98ac4d950e6f0dcf998a8e7dc007e1df8d1553225d" Mar 17 09:26:59 crc kubenswrapper[4813]: I0317 09:26:59.588292 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"723058636df9ee0769fe5a98ac4d950e6f0dcf998a8e7dc007e1df8d1553225d"} err="failed to get container status \"723058636df9ee0769fe5a98ac4d950e6f0dcf998a8e7dc007e1df8d1553225d\": rpc error: code = NotFound desc = could not find container \"723058636df9ee0769fe5a98ac4d950e6f0dcf998a8e7dc007e1df8d1553225d\": container with ID starting with 723058636df9ee0769fe5a98ac4d950e6f0dcf998a8e7dc007e1df8d1553225d not found: ID does not exist" Mar 17 09:27:00 crc kubenswrapper[4813]: I0317 09:27:00.743710 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="733df085-2100-4070-b18e-46ef5390a164" path="/var/lib/kubelet/pods/733df085-2100-4070-b18e-46ef5390a164/volumes" Mar 17 09:27:02 crc kubenswrapper[4813]: I0317 09:27:02.332680 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-z8wcr" Mar 17 09:27:02 crc kubenswrapper[4813]: I0317 09:27:02.332965 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-z8wcr" Mar 17 09:27:02 crc kubenswrapper[4813]: I0317 09:27:02.368323 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-z8wcr" Mar 17 09:27:02 crc kubenswrapper[4813]: I0317 09:27:02.568438 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-z8wcr" Mar 17 09:27:04 crc kubenswrapper[4813]: I0317 09:27:04.231749 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd"] Mar 17 09:27:04 crc kubenswrapper[4813]: E0317 09:27:04.232014 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="733df085-2100-4070-b18e-46ef5390a164" containerName="extract-utilities" Mar 17 09:27:04 crc kubenswrapper[4813]: I0317 09:27:04.232028 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="733df085-2100-4070-b18e-46ef5390a164" containerName="extract-utilities" Mar 17 09:27:04 crc kubenswrapper[4813]: E0317 09:27:04.232042 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="733df085-2100-4070-b18e-46ef5390a164" containerName="registry-server" Mar 17 09:27:04 crc kubenswrapper[4813]: I0317 09:27:04.232051 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="733df085-2100-4070-b18e-46ef5390a164" containerName="registry-server" Mar 17 09:27:04 crc kubenswrapper[4813]: E0317 09:27:04.232066 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="733df085-2100-4070-b18e-46ef5390a164" containerName="extract-content" Mar 17 09:27:04 crc kubenswrapper[4813]: I0317 09:27:04.232075 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="733df085-2100-4070-b18e-46ef5390a164" containerName="extract-content" Mar 17 09:27:04 crc kubenswrapper[4813]: I0317 09:27:04.232240 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="733df085-2100-4070-b18e-46ef5390a164" containerName="registry-server" Mar 17 09:27:04 crc kubenswrapper[4813]: I0317 09:27:04.233214 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd" Mar 17 09:27:04 crc kubenswrapper[4813]: I0317 09:27:04.235711 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-gn5jn" Mar 17 09:27:04 crc kubenswrapper[4813]: I0317 09:27:04.245384 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd"] Mar 17 09:27:04 crc kubenswrapper[4813]: I0317 09:27:04.283483 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fa9ca0da-b53a-42bc-ace9-49579d21ead8-bundle\") pod \"47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd\" (UID: \"fa9ca0da-b53a-42bc-ace9-49579d21ead8\") " pod="openstack-operators/47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd" Mar 17 09:27:04 crc kubenswrapper[4813]: I0317 09:27:04.283636 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckf8j\" (UniqueName: \"kubernetes.io/projected/fa9ca0da-b53a-42bc-ace9-49579d21ead8-kube-api-access-ckf8j\") pod \"47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd\" (UID: \"fa9ca0da-b53a-42bc-ace9-49579d21ead8\") " pod="openstack-operators/47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd" Mar 17 09:27:04 crc kubenswrapper[4813]: I0317 09:27:04.283761 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fa9ca0da-b53a-42bc-ace9-49579d21ead8-util\") pod \"47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd\" (UID: \"fa9ca0da-b53a-42bc-ace9-49579d21ead8\") " pod="openstack-operators/47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd" Mar 17 09:27:04 crc kubenswrapper[4813]: I0317 09:27:04.384784 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fa9ca0da-b53a-42bc-ace9-49579d21ead8-bundle\") pod \"47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd\" (UID: \"fa9ca0da-b53a-42bc-ace9-49579d21ead8\") " pod="openstack-operators/47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd" Mar 17 09:27:04 crc kubenswrapper[4813]: I0317 09:27:04.384869 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckf8j\" (UniqueName: \"kubernetes.io/projected/fa9ca0da-b53a-42bc-ace9-49579d21ead8-kube-api-access-ckf8j\") pod \"47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd\" (UID: \"fa9ca0da-b53a-42bc-ace9-49579d21ead8\") " pod="openstack-operators/47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd" Mar 17 09:27:04 crc kubenswrapper[4813]: I0317 09:27:04.384915 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fa9ca0da-b53a-42bc-ace9-49579d21ead8-util\") pod \"47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd\" (UID: \"fa9ca0da-b53a-42bc-ace9-49579d21ead8\") " pod="openstack-operators/47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd" Mar 17 09:27:04 crc kubenswrapper[4813]: I0317 09:27:04.385717 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fa9ca0da-b53a-42bc-ace9-49579d21ead8-util\") pod \"47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd\" (UID: \"fa9ca0da-b53a-42bc-ace9-49579d21ead8\") " pod="openstack-operators/47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd" Mar 17 09:27:04 crc kubenswrapper[4813]: I0317 09:27:04.385829 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fa9ca0da-b53a-42bc-ace9-49579d21ead8-bundle\") pod \"47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd\" (UID: \"fa9ca0da-b53a-42bc-ace9-49579d21ead8\") " pod="openstack-operators/47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd" Mar 17 09:27:04 crc kubenswrapper[4813]: I0317 09:27:04.418489 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckf8j\" (UniqueName: \"kubernetes.io/projected/fa9ca0da-b53a-42bc-ace9-49579d21ead8-kube-api-access-ckf8j\") pod \"47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd\" (UID: \"fa9ca0da-b53a-42bc-ace9-49579d21ead8\") " pod="openstack-operators/47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd" Mar 17 09:27:04 crc kubenswrapper[4813]: I0317 09:27:04.565024 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd" Mar 17 09:27:04 crc kubenswrapper[4813]: I0317 09:27:04.868181 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd"] Mar 17 09:27:05 crc kubenswrapper[4813]: I0317 09:27:05.556994 4813 generic.go:334] "Generic (PLEG): container finished" podID="fa9ca0da-b53a-42bc-ace9-49579d21ead8" containerID="3e73f15c7c00a8a83813e19ac322fc022033f6bbfddb0f3a3cc9598bed1c6a59" exitCode=0 Mar 17 09:27:05 crc kubenswrapper[4813]: I0317 09:27:05.557067 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd" event={"ID":"fa9ca0da-b53a-42bc-ace9-49579d21ead8","Type":"ContainerDied","Data":"3e73f15c7c00a8a83813e19ac322fc022033f6bbfddb0f3a3cc9598bed1c6a59"} Mar 17 09:27:05 crc kubenswrapper[4813]: I0317 09:27:05.557120 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd" event={"ID":"fa9ca0da-b53a-42bc-ace9-49579d21ead8","Type":"ContainerStarted","Data":"8910ca657390dc7c5a0c63531617da9897865c8e92f5805ebf531167d2bd0d7b"} Mar 17 09:27:06 crc kubenswrapper[4813]: I0317 09:27:06.569571 4813 generic.go:334] "Generic (PLEG): container finished" podID="fa9ca0da-b53a-42bc-ace9-49579d21ead8" containerID="8f4f49777c953eaa01e4f5a15298f7471e3618d2824a042bcbd49ee2ae8214b7" exitCode=0 Mar 17 09:27:06 crc kubenswrapper[4813]: I0317 09:27:06.569664 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd" event={"ID":"fa9ca0da-b53a-42bc-ace9-49579d21ead8","Type":"ContainerDied","Data":"8f4f49777c953eaa01e4f5a15298f7471e3618d2824a042bcbd49ee2ae8214b7"} Mar 17 09:27:07 crc kubenswrapper[4813]: I0317 09:27:07.581099 4813 generic.go:334] "Generic (PLEG): container finished" podID="fa9ca0da-b53a-42bc-ace9-49579d21ead8" containerID="a607e3079c4cb8356231dd54ed5a7120c1f3a3fb39a469719b6275c9894478a0" exitCode=0 Mar 17 09:27:07 crc kubenswrapper[4813]: I0317 09:27:07.581270 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd" event={"ID":"fa9ca0da-b53a-42bc-ace9-49579d21ead8","Type":"ContainerDied","Data":"a607e3079c4cb8356231dd54ed5a7120c1f3a3fb39a469719b6275c9894478a0"} Mar 17 09:27:08 crc kubenswrapper[4813]: I0317 09:27:08.896297 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd" Mar 17 09:27:08 crc kubenswrapper[4813]: I0317 09:27:08.949156 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fa9ca0da-b53a-42bc-ace9-49579d21ead8-bundle\") pod \"fa9ca0da-b53a-42bc-ace9-49579d21ead8\" (UID: \"fa9ca0da-b53a-42bc-ace9-49579d21ead8\") " Mar 17 09:27:08 crc kubenswrapper[4813]: I0317 09:27:08.949224 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fa9ca0da-b53a-42bc-ace9-49579d21ead8-util\") pod \"fa9ca0da-b53a-42bc-ace9-49579d21ead8\" (UID: \"fa9ca0da-b53a-42bc-ace9-49579d21ead8\") " Mar 17 09:27:08 crc kubenswrapper[4813]: I0317 09:27:08.949256 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ckf8j\" (UniqueName: \"kubernetes.io/projected/fa9ca0da-b53a-42bc-ace9-49579d21ead8-kube-api-access-ckf8j\") pod \"fa9ca0da-b53a-42bc-ace9-49579d21ead8\" (UID: \"fa9ca0da-b53a-42bc-ace9-49579d21ead8\") " Mar 17 09:27:08 crc kubenswrapper[4813]: I0317 09:27:08.950184 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa9ca0da-b53a-42bc-ace9-49579d21ead8-bundle" (OuterVolumeSpecName: "bundle") pod "fa9ca0da-b53a-42bc-ace9-49579d21ead8" (UID: "fa9ca0da-b53a-42bc-ace9-49579d21ead8"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:27:08 crc kubenswrapper[4813]: I0317 09:27:08.953987 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa9ca0da-b53a-42bc-ace9-49579d21ead8-kube-api-access-ckf8j" (OuterVolumeSpecName: "kube-api-access-ckf8j") pod "fa9ca0da-b53a-42bc-ace9-49579d21ead8" (UID: "fa9ca0da-b53a-42bc-ace9-49579d21ead8"). InnerVolumeSpecName "kube-api-access-ckf8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:27:08 crc kubenswrapper[4813]: I0317 09:27:08.969131 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fa9ca0da-b53a-42bc-ace9-49579d21ead8-util" (OuterVolumeSpecName: "util") pod "fa9ca0da-b53a-42bc-ace9-49579d21ead8" (UID: "fa9ca0da-b53a-42bc-ace9-49579d21ead8"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:27:09 crc kubenswrapper[4813]: I0317 09:27:09.051383 4813 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/fa9ca0da-b53a-42bc-ace9-49579d21ead8-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:27:09 crc kubenswrapper[4813]: I0317 09:27:09.051441 4813 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/fa9ca0da-b53a-42bc-ace9-49579d21ead8-util\") on node \"crc\" DevicePath \"\"" Mar 17 09:27:09 crc kubenswrapper[4813]: I0317 09:27:09.051469 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ckf8j\" (UniqueName: \"kubernetes.io/projected/fa9ca0da-b53a-42bc-ace9-49579d21ead8-kube-api-access-ckf8j\") on node \"crc\" DevicePath \"\"" Mar 17 09:27:09 crc kubenswrapper[4813]: I0317 09:27:09.600903 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd" event={"ID":"fa9ca0da-b53a-42bc-ace9-49579d21ead8","Type":"ContainerDied","Data":"8910ca657390dc7c5a0c63531617da9897865c8e92f5805ebf531167d2bd0d7b"} Mar 17 09:27:09 crc kubenswrapper[4813]: I0317 09:27:09.600943 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8910ca657390dc7c5a0c63531617da9897865c8e92f5805ebf531167d2bd0d7b" Mar 17 09:27:09 crc kubenswrapper[4813]: I0317 09:27:09.600982 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd" Mar 17 09:27:13 crc kubenswrapper[4813]: I0317 09:27:13.394985 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-init-56ddf7fcb8-7hkdj"] Mar 17 09:27:13 crc kubenswrapper[4813]: E0317 09:27:13.395431 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa9ca0da-b53a-42bc-ace9-49579d21ead8" containerName="util" Mar 17 09:27:13 crc kubenswrapper[4813]: I0317 09:27:13.395441 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa9ca0da-b53a-42bc-ace9-49579d21ead8" containerName="util" Mar 17 09:27:13 crc kubenswrapper[4813]: E0317 09:27:13.395453 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa9ca0da-b53a-42bc-ace9-49579d21ead8" containerName="pull" Mar 17 09:27:13 crc kubenswrapper[4813]: I0317 09:27:13.395458 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa9ca0da-b53a-42bc-ace9-49579d21ead8" containerName="pull" Mar 17 09:27:13 crc kubenswrapper[4813]: E0317 09:27:13.395479 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa9ca0da-b53a-42bc-ace9-49579d21ead8" containerName="extract" Mar 17 09:27:13 crc kubenswrapper[4813]: I0317 09:27:13.395485 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa9ca0da-b53a-42bc-ace9-49579d21ead8" containerName="extract" Mar 17 09:27:13 crc kubenswrapper[4813]: I0317 09:27:13.395581 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa9ca0da-b53a-42bc-ace9-49579d21ead8" containerName="extract" Mar 17 09:27:13 crc kubenswrapper[4813]: I0317 09:27:13.395956 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-56ddf7fcb8-7hkdj" Mar 17 09:27:13 crc kubenswrapper[4813]: I0317 09:27:13.398228 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-init-dockercfg-hvgrg" Mar 17 09:27:13 crc kubenswrapper[4813]: I0317 09:27:13.425533 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-56ddf7fcb8-7hkdj"] Mar 17 09:27:13 crc kubenswrapper[4813]: I0317 09:27:13.512776 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sg7xf\" (UniqueName: \"kubernetes.io/projected/6dc9eebf-39dd-4030-a091-5626832fd141-kube-api-access-sg7xf\") pod \"openstack-operator-controller-init-56ddf7fcb8-7hkdj\" (UID: \"6dc9eebf-39dd-4030-a091-5626832fd141\") " pod="openstack-operators/openstack-operator-controller-init-56ddf7fcb8-7hkdj" Mar 17 09:27:13 crc kubenswrapper[4813]: I0317 09:27:13.614047 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sg7xf\" (UniqueName: \"kubernetes.io/projected/6dc9eebf-39dd-4030-a091-5626832fd141-kube-api-access-sg7xf\") pod \"openstack-operator-controller-init-56ddf7fcb8-7hkdj\" (UID: \"6dc9eebf-39dd-4030-a091-5626832fd141\") " pod="openstack-operators/openstack-operator-controller-init-56ddf7fcb8-7hkdj" Mar 17 09:27:13 crc kubenswrapper[4813]: I0317 09:27:13.633791 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sg7xf\" (UniqueName: \"kubernetes.io/projected/6dc9eebf-39dd-4030-a091-5626832fd141-kube-api-access-sg7xf\") pod \"openstack-operator-controller-init-56ddf7fcb8-7hkdj\" (UID: \"6dc9eebf-39dd-4030-a091-5626832fd141\") " pod="openstack-operators/openstack-operator-controller-init-56ddf7fcb8-7hkdj" Mar 17 09:27:13 crc kubenswrapper[4813]: I0317 09:27:13.715237 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-init-56ddf7fcb8-7hkdj" Mar 17 09:27:14 crc kubenswrapper[4813]: I0317 09:27:14.040282 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-init-56ddf7fcb8-7hkdj"] Mar 17 09:27:14 crc kubenswrapper[4813]: I0317 09:27:14.115767 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:27:14 crc kubenswrapper[4813]: I0317 09:27:14.115835 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:27:14 crc kubenswrapper[4813]: I0317 09:27:14.115887 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:27:14 crc kubenswrapper[4813]: I0317 09:27:14.116645 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c178911cb0823cd7d042791f6122482441f05cc74687451c72a3d02c4974d8e8"} pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 09:27:14 crc kubenswrapper[4813]: I0317 09:27:14.116704 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" containerID="cri-o://c178911cb0823cd7d042791f6122482441f05cc74687451c72a3d02c4974d8e8" gracePeriod=600 Mar 17 09:27:14 crc kubenswrapper[4813]: I0317 09:27:14.640578 4813 generic.go:334] "Generic (PLEG): container finished" podID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerID="c178911cb0823cd7d042791f6122482441f05cc74687451c72a3d02c4974d8e8" exitCode=0 Mar 17 09:27:14 crc kubenswrapper[4813]: I0317 09:27:14.640932 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerDied","Data":"c178911cb0823cd7d042791f6122482441f05cc74687451c72a3d02c4974d8e8"} Mar 17 09:27:14 crc kubenswrapper[4813]: I0317 09:27:14.640961 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerStarted","Data":"d74d47d3497863468b9c4543e558eb883310048990c8823420967f1910298e01"} Mar 17 09:27:14 crc kubenswrapper[4813]: I0317 09:27:14.640980 4813 scope.go:117] "RemoveContainer" containerID="b67bb2399345967e5d98c3e93bfb13d1f15014e04cd83984dc098f8a87b2a52c" Mar 17 09:27:14 crc kubenswrapper[4813]: I0317 09:27:14.643430 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-56ddf7fcb8-7hkdj" event={"ID":"6dc9eebf-39dd-4030-a091-5626832fd141","Type":"ContainerStarted","Data":"01ef4e9c37dca3cc47f1545d4d405adba393ffb6f0a8b6b7bc49e38d10bfb9a4"} Mar 17 09:27:18 crc kubenswrapper[4813]: I0317 09:27:18.680193 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-init-56ddf7fcb8-7hkdj" event={"ID":"6dc9eebf-39dd-4030-a091-5626832fd141","Type":"ContainerStarted","Data":"02f006e224e69d3552528ee13fd6fc0c82e421e48ca31e876b87b87d9531805d"} Mar 17 09:27:18 crc kubenswrapper[4813]: I0317 09:27:18.680883 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-init-56ddf7fcb8-7hkdj" Mar 17 09:27:18 crc kubenswrapper[4813]: I0317 09:27:18.725295 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-init-56ddf7fcb8-7hkdj" podStartSLOduration=1.865678927 podStartE2EDuration="5.725264688s" podCreationTimestamp="2026-03-17 09:27:13 +0000 UTC" firstStartedPulling="2026-03-17 09:27:14.052577638 +0000 UTC m=+1056.153381157" lastFinishedPulling="2026-03-17 09:27:17.912163419 +0000 UTC m=+1060.012966918" observedRunningTime="2026-03-17 09:27:18.712643461 +0000 UTC m=+1060.813446970" watchObservedRunningTime="2026-03-17 09:27:18.725264688 +0000 UTC m=+1060.826068237" Mar 17 09:27:23 crc kubenswrapper[4813]: I0317 09:27:23.718455 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-init-56ddf7fcb8-7hkdj" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.664082 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-xkfkm"] Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.665500 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-xkfkm" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.672036 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-bhkc6" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.682304 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-9cqth"] Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.683067 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-9cqth" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.684703 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-dq764" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.691047 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-xkfkm"] Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.697755 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-9cqth"] Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.709239 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-tn4pv"] Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.709960 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-tn4pv" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.716795 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-cvfxp" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.727708 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-wzmxx"] Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.728541 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-wzmxx" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.734501 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-x68ld" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.749807 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-tn4pv"] Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.765481 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-wzmxx"] Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.772403 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76xbr\" (UniqueName: \"kubernetes.io/projected/9b3cf477-8c47-4a38-beee-8bd3f70977e0-kube-api-access-76xbr\") pod \"designate-operator-controller-manager-588d4d986b-tn4pv\" (UID: \"9b3cf477-8c47-4a38-beee-8bd3f70977e0\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-tn4pv" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.772451 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nxsxn\" (UniqueName: \"kubernetes.io/projected/227f29cf-ec7d-40e9-b28c-c1ccae9d880e-kube-api-access-nxsxn\") pod \"barbican-operator-controller-manager-59bc569d95-xkfkm\" (UID: \"227f29cf-ec7d-40e9-b28c-c1ccae9d880e\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-xkfkm" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.772486 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57rlx\" (UniqueName: \"kubernetes.io/projected/1b2b8088-31e5-49af-91b4-a3fbfe797af5-kube-api-access-57rlx\") pod \"cinder-operator-controller-manager-8d58dc466-9cqth\" (UID: \"1b2b8088-31e5-49af-91b4-a3fbfe797af5\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-9cqth" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.777078 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-hng2s"] Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.778207 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-hng2s" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.783022 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-qnjzn" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.792449 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-hng2s"] Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.847267 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-tfhsc"] Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.855399 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-tfhsc" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.858961 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-mlhkf" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.872481 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-7b9c774f96-7965b"] Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.878025 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-7965b" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.878396 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrrsm\" (UniqueName: \"kubernetes.io/projected/de6c9679-9f9f-481d-85d9-218c9199088e-kube-api-access-mrrsm\") pod \"heat-operator-controller-manager-67dd5f86f5-hng2s\" (UID: \"de6c9679-9f9f-481d-85d9-218c9199088e\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-hng2s" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.879696 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzcdf\" (UniqueName: \"kubernetes.io/projected/3c5a1075-8105-401b-afb0-00014dafed0a-kube-api-access-fzcdf\") pod \"glance-operator-controller-manager-79df6bcc97-wzmxx\" (UID: \"3c5a1075-8105-401b-afb0-00014dafed0a\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-wzmxx" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.879984 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76xbr\" (UniqueName: \"kubernetes.io/projected/9b3cf477-8c47-4a38-beee-8bd3f70977e0-kube-api-access-76xbr\") pod \"designate-operator-controller-manager-588d4d986b-tn4pv\" (UID: \"9b3cf477-8c47-4a38-beee-8bd3f70977e0\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-tn4pv" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.880276 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nxsxn\" (UniqueName: \"kubernetes.io/projected/227f29cf-ec7d-40e9-b28c-c1ccae9d880e-kube-api-access-nxsxn\") pod \"barbican-operator-controller-manager-59bc569d95-xkfkm\" (UID: \"227f29cf-ec7d-40e9-b28c-c1ccae9d880e\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-xkfkm" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.880748 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.880895 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57rlx\" (UniqueName: \"kubernetes.io/projected/1b2b8088-31e5-49af-91b4-a3fbfe797af5-kube-api-access-57rlx\") pod \"cinder-operator-controller-manager-8d58dc466-9cqth\" (UID: \"1b2b8088-31e5-49af-91b4-a3fbfe797af5\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-9cqth" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.882331 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-rt7sh" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.888742 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-tfhsc"] Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.902693 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7b9c774f96-7965b"] Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.911802 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nxsxn\" (UniqueName: \"kubernetes.io/projected/227f29cf-ec7d-40e9-b28c-c1ccae9d880e-kube-api-access-nxsxn\") pod \"barbican-operator-controller-manager-59bc569d95-xkfkm\" (UID: \"227f29cf-ec7d-40e9-b28c-c1ccae9d880e\") " pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-xkfkm" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.914390 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-4ds8c"] Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.915199 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-4ds8c" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.932639 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-c966n"] Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.933424 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c966n" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.945746 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-4ds8c"] Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.946365 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-jwqh8" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.953822 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-2prpw" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.971788 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57rlx\" (UniqueName: \"kubernetes.io/projected/1b2b8088-31e5-49af-91b4-a3fbfe797af5-kube-api-access-57rlx\") pod \"cinder-operator-controller-manager-8d58dc466-9cqth\" (UID: \"1b2b8088-31e5-49af-91b4-a3fbfe797af5\") " pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-9cqth" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.977228 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76xbr\" (UniqueName: \"kubernetes.io/projected/9b3cf477-8c47-4a38-beee-8bd3f70977e0-kube-api-access-76xbr\") pod \"designate-operator-controller-manager-588d4d986b-tn4pv\" (UID: \"9b3cf477-8c47-4a38-beee-8bd3f70977e0\") " pod="openstack-operators/designate-operator-controller-manager-588d4d986b-tn4pv" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.993615 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-dbs68"] Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.994431 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-55f864c847-dbs68" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.996255 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jd62n\" (UniqueName: \"kubernetes.io/projected/fd3396be-4825-4edc-a41f-443af1413e3b-kube-api-access-jd62n\") pod \"keystone-operator-controller-manager-768b96df4c-c966n\" (UID: \"fd3396be-4825-4edc-a41f-443af1413e3b\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c966n" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.996318 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrrsm\" (UniqueName: \"kubernetes.io/projected/de6c9679-9f9f-481d-85d9-218c9199088e-kube-api-access-mrrsm\") pod \"heat-operator-controller-manager-67dd5f86f5-hng2s\" (UID: \"de6c9679-9f9f-481d-85d9-218c9199088e\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-hng2s" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.996343 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqh9p\" (UniqueName: \"kubernetes.io/projected/c9ea4c59-d87b-4e58-ac8f-2786f8f32f02-kube-api-access-pqh9p\") pod \"infra-operator-controller-manager-7b9c774f96-7965b\" (UID: \"c9ea4c59-d87b-4e58-ac8f-2786f8f32f02\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-7965b" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.996364 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzcdf\" (UniqueName: \"kubernetes.io/projected/3c5a1075-8105-401b-afb0-00014dafed0a-kube-api-access-fzcdf\") pod \"glance-operator-controller-manager-79df6bcc97-wzmxx\" (UID: \"3c5a1075-8105-401b-afb0-00014dafed0a\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-wzmxx" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.996386 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c9ea4c59-d87b-4e58-ac8f-2786f8f32f02-cert\") pod \"infra-operator-controller-manager-7b9c774f96-7965b\" (UID: \"c9ea4c59-d87b-4e58-ac8f-2786f8f32f02\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-7965b" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.996412 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6868\" (UniqueName: \"kubernetes.io/projected/668691cc-39f3-4b04-8ed6-bc86f9965f12-kube-api-access-d6868\") pod \"horizon-operator-controller-manager-8464cc45fb-tfhsc\" (UID: \"668691cc-39f3-4b04-8ed6-bc86f9965f12\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-tfhsc" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.996481 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzhsh\" (UniqueName: \"kubernetes.io/projected/80f14d5d-c8b9-4068-9469-ea8eff96e9e5-kube-api-access-vzhsh\") pod \"ironic-operator-controller-manager-6f787dddc9-4ds8c\" (UID: \"80f14d5d-c8b9-4068-9469-ea8eff96e9e5\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-4ds8c" Mar 17 09:27:43 crc kubenswrapper[4813]: I0317 09:27:43.996941 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-xkfkm" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.008153 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-9cqth" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.009169 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-f8stl" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.020267 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-c966n"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.031137 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-tn4pv" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.032786 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzcdf\" (UniqueName: \"kubernetes.io/projected/3c5a1075-8105-401b-afb0-00014dafed0a-kube-api-access-fzcdf\") pod \"glance-operator-controller-manager-79df6bcc97-wzmxx\" (UID: \"3c5a1075-8105-401b-afb0-00014dafed0a\") " pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-wzmxx" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.042778 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-dbs68"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.046958 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-wzmxx" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.048227 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrrsm\" (UniqueName: \"kubernetes.io/projected/de6c9679-9f9f-481d-85d9-218c9199088e-kube-api-access-mrrsm\") pod \"heat-operator-controller-manager-67dd5f86f5-hng2s\" (UID: \"de6c9679-9f9f-481d-85d9-218c9199088e\") " pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-hng2s" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.058764 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ngd7n"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.059940 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ngd7n" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.077006 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-489gn" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.096923 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-mnsb4"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.097525 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzhsh\" (UniqueName: \"kubernetes.io/projected/80f14d5d-c8b9-4068-9469-ea8eff96e9e5-kube-api-access-vzhsh\") pod \"ironic-operator-controller-manager-6f787dddc9-4ds8c\" (UID: \"80f14d5d-c8b9-4068-9469-ea8eff96e9e5\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-4ds8c" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.097563 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jd62n\" (UniqueName: \"kubernetes.io/projected/fd3396be-4825-4edc-a41f-443af1413e3b-kube-api-access-jd62n\") pod \"keystone-operator-controller-manager-768b96df4c-c966n\" (UID: \"fd3396be-4825-4edc-a41f-443af1413e3b\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c966n" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.097609 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qzrs\" (UniqueName: \"kubernetes.io/projected/5e7b137d-9b58-448b-91a4-3a069dfb4d10-kube-api-access-2qzrs\") pod \"manila-operator-controller-manager-55f864c847-dbs68\" (UID: \"5e7b137d-9b58-448b-91a4-3a069dfb4d10\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-dbs68" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.097632 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nrf9m\" (UniqueName: \"kubernetes.io/projected/0e3eba88-a4a4-4c6d-9e90-5dba41268494-kube-api-access-nrf9m\") pod \"mariadb-operator-controller-manager-67ccfc9778-ngd7n\" (UID: \"0e3eba88-a4a4-4c6d-9e90-5dba41268494\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ngd7n" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.097656 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqh9p\" (UniqueName: \"kubernetes.io/projected/c9ea4c59-d87b-4e58-ac8f-2786f8f32f02-kube-api-access-pqh9p\") pod \"infra-operator-controller-manager-7b9c774f96-7965b\" (UID: \"c9ea4c59-d87b-4e58-ac8f-2786f8f32f02\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-7965b" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.097681 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c9ea4c59-d87b-4e58-ac8f-2786f8f32f02-cert\") pod \"infra-operator-controller-manager-7b9c774f96-7965b\" (UID: \"c9ea4c59-d87b-4e58-ac8f-2786f8f32f02\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-7965b" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.097705 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6868\" (UniqueName: \"kubernetes.io/projected/668691cc-39f3-4b04-8ed6-bc86f9965f12-kube-api-access-d6868\") pod \"horizon-operator-controller-manager-8464cc45fb-tfhsc\" (UID: \"668691cc-39f3-4b04-8ed6-bc86f9965f12\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-tfhsc" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.097800 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-767865f676-mnsb4" Mar 17 09:27:44 crc kubenswrapper[4813]: E0317 09:27:44.098208 4813 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 17 09:27:44 crc kubenswrapper[4813]: E0317 09:27:44.098254 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9ea4c59-d87b-4e58-ac8f-2786f8f32f02-cert podName:c9ea4c59-d87b-4e58-ac8f-2786f8f32f02 nodeName:}" failed. No retries permitted until 2026-03-17 09:27:44.598239671 +0000 UTC m=+1086.699043170 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c9ea4c59-d87b-4e58-ac8f-2786f8f32f02-cert") pod "infra-operator-controller-manager-7b9c774f96-7965b" (UID: "c9ea4c59-d87b-4e58-ac8f-2786f8f32f02") : secret "infra-operator-webhook-server-cert" not found Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.101111 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-sqr76" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.107306 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ngd7n"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.163476 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6868\" (UniqueName: \"kubernetes.io/projected/668691cc-39f3-4b04-8ed6-bc86f9965f12-kube-api-access-d6868\") pod \"horizon-operator-controller-manager-8464cc45fb-tfhsc\" (UID: \"668691cc-39f3-4b04-8ed6-bc86f9965f12\") " pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-tfhsc" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.163563 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-mnsb4"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.164866 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzhsh\" (UniqueName: \"kubernetes.io/projected/80f14d5d-c8b9-4068-9469-ea8eff96e9e5-kube-api-access-vzhsh\") pod \"ironic-operator-controller-manager-6f787dddc9-4ds8c\" (UID: \"80f14d5d-c8b9-4068-9469-ea8eff96e9e5\") " pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-4ds8c" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.181551 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jd62n\" (UniqueName: \"kubernetes.io/projected/fd3396be-4825-4edc-a41f-443af1413e3b-kube-api-access-jd62n\") pod \"keystone-operator-controller-manager-768b96df4c-c966n\" (UID: \"fd3396be-4825-4edc-a41f-443af1413e3b\") " pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c966n" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.184877 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqh9p\" (UniqueName: \"kubernetes.io/projected/c9ea4c59-d87b-4e58-ac8f-2786f8f32f02-kube-api-access-pqh9p\") pod \"infra-operator-controller-manager-7b9c774f96-7965b\" (UID: \"c9ea4c59-d87b-4e58-ac8f-2786f8f32f02\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-7965b" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.190094 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-hng2s" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.195005 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-tqg57"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.195796 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tqg57" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.199364 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qzrs\" (UniqueName: \"kubernetes.io/projected/5e7b137d-9b58-448b-91a4-3a069dfb4d10-kube-api-access-2qzrs\") pod \"manila-operator-controller-manager-55f864c847-dbs68\" (UID: \"5e7b137d-9b58-448b-91a4-3a069dfb4d10\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-dbs68" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.199402 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nrf9m\" (UniqueName: \"kubernetes.io/projected/0e3eba88-a4a4-4c6d-9e90-5dba41268494-kube-api-access-nrf9m\") pod \"mariadb-operator-controller-manager-67ccfc9778-ngd7n\" (UID: \"0e3eba88-a4a4-4c6d-9e90-5dba41268494\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ngd7n" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.199423 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92bd4\" (UniqueName: \"kubernetes.io/projected/c3419efc-507c-46bc-aedb-ef5eb3f0f917-kube-api-access-92bd4\") pod \"neutron-operator-controller-manager-767865f676-mnsb4\" (UID: \"c3419efc-507c-46bc-aedb-ef5eb3f0f917\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-mnsb4" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.203073 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-27t8q" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.211644 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-mps6t"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.212578 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-mps6t" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.231466 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-47k4q" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.235507 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-tqg57"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.259968 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qzrs\" (UniqueName: \"kubernetes.io/projected/5e7b137d-9b58-448b-91a4-3a069dfb4d10-kube-api-access-2qzrs\") pod \"manila-operator-controller-manager-55f864c847-dbs68\" (UID: \"5e7b137d-9b58-448b-91a4-3a069dfb4d10\") " pod="openstack-operators/manila-operator-controller-manager-55f864c847-dbs68" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.262684 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-mps6t"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.265745 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nrf9m\" (UniqueName: \"kubernetes.io/projected/0e3eba88-a4a4-4c6d-9e90-5dba41268494-kube-api-access-nrf9m\") pod \"mariadb-operator-controller-manager-67ccfc9778-ngd7n\" (UID: \"0e3eba88-a4a4-4c6d-9e90-5dba41268494\") " pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ngd7n" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.294431 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ngd7n" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.301609 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92bd4\" (UniqueName: \"kubernetes.io/projected/c3419efc-507c-46bc-aedb-ef5eb3f0f917-kube-api-access-92bd4\") pod \"neutron-operator-controller-manager-767865f676-mnsb4\" (UID: \"c3419efc-507c-46bc-aedb-ef5eb3f0f917\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-mnsb4" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.301685 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfqtl\" (UniqueName: \"kubernetes.io/projected/b4b9a61e-819f-4da5-a28a-8e49c0a3f7cd-kube-api-access-bfqtl\") pod \"nova-operator-controller-manager-5d488d59fb-tqg57\" (UID: \"b4b9a61e-819f-4da5-a28a-8e49c0a3f7cd\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tqg57" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.301780 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cqvkd\" (UniqueName: \"kubernetes.io/projected/c6e1a616-6d48-4ef1-93cb-62caed2de963-kube-api-access-cqvkd\") pod \"octavia-operator-controller-manager-5b9f45d989-mps6t\" (UID: \"c6e1a616-6d48-4ef1-93cb-62caed2de963\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-mps6t" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.307175 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-hcsvd"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.307953 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5784578c99-hcsvd" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.313333 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-fmd7w" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.313533 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-7d66d"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.314311 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-884679f54-7d66d" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.316081 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-7ccsj" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.326695 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-hcsvd"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.337943 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.338721 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.344667 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-fh4tx" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.345161 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.349251 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92bd4\" (UniqueName: \"kubernetes.io/projected/c3419efc-507c-46bc-aedb-ef5eb3f0f917-kube-api-access-92bd4\") pod \"neutron-operator-controller-manager-767865f676-mnsb4\" (UID: \"c3419efc-507c-46bc-aedb-ef5eb3f0f917\") " pod="openstack-operators/neutron-operator-controller-manager-767865f676-mnsb4" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.356058 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-njxx7"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.357227 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-c674c5965-njxx7" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.365538 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-84sv8" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.376434 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-7d66d"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.377909 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c966n" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.382352 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-52xx2"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.383757 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-52xx2" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.386444 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-fm45k" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.390121 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-r2ksk"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.392339 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-r2ksk" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.394547 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-594wz" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.446750 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqvkd\" (UniqueName: \"kubernetes.io/projected/c6e1a616-6d48-4ef1-93cb-62caed2de963-kube-api-access-cqvkd\") pod \"octavia-operator-controller-manager-5b9f45d989-mps6t\" (UID: \"c6e1a616-6d48-4ef1-93cb-62caed2de963\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-mps6t" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.446922 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zthsr\" (UniqueName: \"kubernetes.io/projected/ac052585-f733-47d5-91b9-87ad7957e511-kube-api-access-zthsr\") pod \"swift-operator-controller-manager-c674c5965-njxx7\" (UID: \"ac052585-f733-47d5-91b9-87ad7957e511\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-njxx7" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.447034 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jnfhn\" (UniqueName: \"kubernetes.io/projected/968158c3-2dfc-439e-9d20-8706d611fec1-kube-api-access-jnfhn\") pod \"placement-operator-controller-manager-5784578c99-hcsvd\" (UID: \"968158c3-2dfc-439e-9d20-8706d611fec1\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-hcsvd" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.447119 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8e80dcf7-dae9-4faa-ab3f-b547b570926b-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-g5vgd\" (UID: \"8e80dcf7-dae9-4faa-ab3f-b547b570926b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.447219 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s7m6p\" (UniqueName: \"kubernetes.io/projected/76eb47fe-8f18-4f98-a8ba-d89169d0de93-kube-api-access-s7m6p\") pod \"ovn-operator-controller-manager-884679f54-7d66d\" (UID: \"76eb47fe-8f18-4f98-a8ba-d89169d0de93\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-7d66d" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.447339 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfqtl\" (UniqueName: \"kubernetes.io/projected/b4b9a61e-819f-4da5-a28a-8e49c0a3f7cd-kube-api-access-bfqtl\") pod \"nova-operator-controller-manager-5d488d59fb-tqg57\" (UID: \"b4b9a61e-819f-4da5-a28a-8e49c0a3f7cd\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tqg57" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.447437 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sxdmj\" (UniqueName: \"kubernetes.io/projected/8e80dcf7-dae9-4faa-ab3f-b547b570926b-kube-api-access-sxdmj\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-g5vgd\" (UID: \"8e80dcf7-dae9-4faa-ab3f-b547b570926b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.454135 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-tfhsc" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.459878 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-55f864c847-dbs68" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.494171 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqvkd\" (UniqueName: \"kubernetes.io/projected/c6e1a616-6d48-4ef1-93cb-62caed2de963-kube-api-access-cqvkd\") pod \"octavia-operator-controller-manager-5b9f45d989-mps6t\" (UID: \"c6e1a616-6d48-4ef1-93cb-62caed2de963\") " pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-mps6t" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.505442 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-767865f676-mnsb4" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.519037 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-4ds8c" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.544700 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfqtl\" (UniqueName: \"kubernetes.io/projected/b4b9a61e-819f-4da5-a28a-8e49c0a3f7cd-kube-api-access-bfqtl\") pod \"nova-operator-controller-manager-5d488d59fb-tqg57\" (UID: \"b4b9a61e-819f-4da5-a28a-8e49c0a3f7cd\") " pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tqg57" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.547094 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-njxx7"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.549795 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6wxb\" (UniqueName: \"kubernetes.io/projected/ec48dbf8-730b-4a2e-9986-03d8f66de013-kube-api-access-b6wxb\") pod \"telemetry-operator-controller-manager-d6b694c5-52xx2\" (UID: \"ec48dbf8-730b-4a2e-9986-03d8f66de013\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-52xx2" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.549840 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jnfhn\" (UniqueName: \"kubernetes.io/projected/968158c3-2dfc-439e-9d20-8706d611fec1-kube-api-access-jnfhn\") pod \"placement-operator-controller-manager-5784578c99-hcsvd\" (UID: \"968158c3-2dfc-439e-9d20-8706d611fec1\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-hcsvd" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.549869 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8e80dcf7-dae9-4faa-ab3f-b547b570926b-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-g5vgd\" (UID: \"8e80dcf7-dae9-4faa-ab3f-b547b570926b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.549898 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s7m6p\" (UniqueName: \"kubernetes.io/projected/76eb47fe-8f18-4f98-a8ba-d89169d0de93-kube-api-access-s7m6p\") pod \"ovn-operator-controller-manager-884679f54-7d66d\" (UID: \"76eb47fe-8f18-4f98-a8ba-d89169d0de93\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-7d66d" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.549956 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sxdmj\" (UniqueName: \"kubernetes.io/projected/8e80dcf7-dae9-4faa-ab3f-b547b570926b-kube-api-access-sxdmj\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-g5vgd\" (UID: \"8e80dcf7-dae9-4faa-ab3f-b547b570926b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.550006 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zthsr\" (UniqueName: \"kubernetes.io/projected/ac052585-f733-47d5-91b9-87ad7957e511-kube-api-access-zthsr\") pod \"swift-operator-controller-manager-c674c5965-njxx7\" (UID: \"ac052585-f733-47d5-91b9-87ad7957e511\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-njxx7" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.550027 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zv8d6\" (UniqueName: \"kubernetes.io/projected/860ccadc-ca76-4523-9543-2b0960eb77a9-kube-api-access-zv8d6\") pod \"test-operator-controller-manager-5c5cb9c4d7-r2ksk\" (UID: \"860ccadc-ca76-4523-9543-2b0960eb77a9\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-r2ksk" Mar 17 09:27:44 crc kubenswrapper[4813]: E0317 09:27:44.550360 4813 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 09:27:44 crc kubenswrapper[4813]: E0317 09:27:44.550393 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e80dcf7-dae9-4faa-ab3f-b547b570926b-cert podName:8e80dcf7-dae9-4faa-ab3f-b547b570926b nodeName:}" failed. No retries permitted until 2026-03-17 09:27:45.050381993 +0000 UTC m=+1087.151185492 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8e80dcf7-dae9-4faa-ab3f-b547b570926b-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" (UID: "8e80dcf7-dae9-4faa-ab3f-b547b570926b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.558308 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.568556 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-52xx2"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.574183 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sxdmj\" (UniqueName: \"kubernetes.io/projected/8e80dcf7-dae9-4faa-ab3f-b547b570926b-kube-api-access-sxdmj\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-g5vgd\" (UID: \"8e80dcf7-dae9-4faa-ab3f-b547b570926b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.574339 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jnfhn\" (UniqueName: \"kubernetes.io/projected/968158c3-2dfc-439e-9d20-8706d611fec1-kube-api-access-jnfhn\") pod \"placement-operator-controller-manager-5784578c99-hcsvd\" (UID: \"968158c3-2dfc-439e-9d20-8706d611fec1\") " pod="openstack-operators/placement-operator-controller-manager-5784578c99-hcsvd" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.574938 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s7m6p\" (UniqueName: \"kubernetes.io/projected/76eb47fe-8f18-4f98-a8ba-d89169d0de93-kube-api-access-s7m6p\") pod \"ovn-operator-controller-manager-884679f54-7d66d\" (UID: \"76eb47fe-8f18-4f98-a8ba-d89169d0de93\") " pod="openstack-operators/ovn-operator-controller-manager-884679f54-7d66d" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.575369 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zthsr\" (UniqueName: \"kubernetes.io/projected/ac052585-f733-47d5-91b9-87ad7957e511-kube-api-access-zthsr\") pod \"swift-operator-controller-manager-c674c5965-njxx7\" (UID: \"ac052585-f733-47d5-91b9-87ad7957e511\") " pod="openstack-operators/swift-operator-controller-manager-c674c5965-njxx7" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.580790 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-wrtl9"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.581983 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-wrtl9" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.584804 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-9k48k" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.587541 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-r2ksk"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.591633 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-wrtl9"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.612762 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tqg57" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.633234 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.634792 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.634904 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.640001 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-mps6t" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.645555 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.645766 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.645962 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-95tfz" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.650816 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zv8d6\" (UniqueName: \"kubernetes.io/projected/860ccadc-ca76-4523-9543-2b0960eb77a9-kube-api-access-zv8d6\") pod \"test-operator-controller-manager-5c5cb9c4d7-r2ksk\" (UID: \"860ccadc-ca76-4523-9543-2b0960eb77a9\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-r2ksk" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.650891 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkjbm\" (UniqueName: \"kubernetes.io/projected/520d3ab3-78d7-4be3-8466-9dadfce15c2c-kube-api-access-xkjbm\") pod \"watcher-operator-controller-manager-6c4d75f7f9-wrtl9\" (UID: \"520d3ab3-78d7-4be3-8466-9dadfce15c2c\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-wrtl9" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.650922 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6wxb\" (UniqueName: \"kubernetes.io/projected/ec48dbf8-730b-4a2e-9986-03d8f66de013-kube-api-access-b6wxb\") pod \"telemetry-operator-controller-manager-d6b694c5-52xx2\" (UID: \"ec48dbf8-730b-4a2e-9986-03d8f66de013\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-52xx2" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.651040 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c9ea4c59-d87b-4e58-ac8f-2786f8f32f02-cert\") pod \"infra-operator-controller-manager-7b9c774f96-7965b\" (UID: \"c9ea4c59-d87b-4e58-ac8f-2786f8f32f02\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-7965b" Mar 17 09:27:44 crc kubenswrapper[4813]: E0317 09:27:44.651180 4813 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 17 09:27:44 crc kubenswrapper[4813]: E0317 09:27:44.651278 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9ea4c59-d87b-4e58-ac8f-2786f8f32f02-cert podName:c9ea4c59-d87b-4e58-ac8f-2786f8f32f02 nodeName:}" failed. No retries permitted until 2026-03-17 09:27:45.651255461 +0000 UTC m=+1087.752058960 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c9ea4c59-d87b-4e58-ac8f-2786f8f32f02-cert") pod "infra-operator-controller-manager-7b9c774f96-7965b" (UID: "c9ea4c59-d87b-4e58-ac8f-2786f8f32f02") : secret "infra-operator-webhook-server-cert" not found Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.660588 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-5784578c99-hcsvd" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.676219 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rxzjs"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.678025 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rxzjs" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.686283 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rxzjs"] Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.686336 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-h6cbz" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.698913 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zv8d6\" (UniqueName: \"kubernetes.io/projected/860ccadc-ca76-4523-9543-2b0960eb77a9-kube-api-access-zv8d6\") pod \"test-operator-controller-manager-5c5cb9c4d7-r2ksk\" (UID: \"860ccadc-ca76-4523-9543-2b0960eb77a9\") " pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-r2ksk" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.700080 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6wxb\" (UniqueName: \"kubernetes.io/projected/ec48dbf8-730b-4a2e-9986-03d8f66de013-kube-api-access-b6wxb\") pod \"telemetry-operator-controller-manager-d6b694c5-52xx2\" (UID: \"ec48dbf8-730b-4a2e-9986-03d8f66de013\") " pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-52xx2" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.718515 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-884679f54-7d66d" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.755239 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bkzld\" (UniqueName: \"kubernetes.io/projected/bcd2954f-b240-4cc3-8387-15b0e64a9721-kube-api-access-bkzld\") pod \"rabbitmq-cluster-operator-manager-668c99d594-rxzjs\" (UID: \"bcd2954f-b240-4cc3-8387-15b0e64a9721\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rxzjs" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.755291 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkjbm\" (UniqueName: \"kubernetes.io/projected/520d3ab3-78d7-4be3-8466-9dadfce15c2c-kube-api-access-xkjbm\") pod \"watcher-operator-controller-manager-6c4d75f7f9-wrtl9\" (UID: \"520d3ab3-78d7-4be3-8466-9dadfce15c2c\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-wrtl9" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.755319 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-webhook-certs\") pod \"openstack-operator-controller-manager-65f8859ffd-mxfzh\" (UID: \"db193bcc-748f-4830-873c-48a2f9229b40\") " pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.755381 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-metrics-certs\") pod \"openstack-operator-controller-manager-65f8859ffd-mxfzh\" (UID: \"db193bcc-748f-4830-873c-48a2f9229b40\") " pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.755424 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phmjt\" (UniqueName: \"kubernetes.io/projected/db193bcc-748f-4830-873c-48a2f9229b40-kube-api-access-phmjt\") pod \"openstack-operator-controller-manager-65f8859ffd-mxfzh\" (UID: \"db193bcc-748f-4830-873c-48a2f9229b40\") " pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.781865 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkjbm\" (UniqueName: \"kubernetes.io/projected/520d3ab3-78d7-4be3-8466-9dadfce15c2c-kube-api-access-xkjbm\") pod \"watcher-operator-controller-manager-6c4d75f7f9-wrtl9\" (UID: \"520d3ab3-78d7-4be3-8466-9dadfce15c2c\") " pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-wrtl9" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.784759 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-c674c5965-njxx7" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.804060 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-52xx2" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.834502 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-r2ksk" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.856483 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bkzld\" (UniqueName: \"kubernetes.io/projected/bcd2954f-b240-4cc3-8387-15b0e64a9721-kube-api-access-bkzld\") pod \"rabbitmq-cluster-operator-manager-668c99d594-rxzjs\" (UID: \"bcd2954f-b240-4cc3-8387-15b0e64a9721\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rxzjs" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.856545 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-webhook-certs\") pod \"openstack-operator-controller-manager-65f8859ffd-mxfzh\" (UID: \"db193bcc-748f-4830-873c-48a2f9229b40\") " pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.856646 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-metrics-certs\") pod \"openstack-operator-controller-manager-65f8859ffd-mxfzh\" (UID: \"db193bcc-748f-4830-873c-48a2f9229b40\") " pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.856728 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phmjt\" (UniqueName: \"kubernetes.io/projected/db193bcc-748f-4830-873c-48a2f9229b40-kube-api-access-phmjt\") pod \"openstack-operator-controller-manager-65f8859ffd-mxfzh\" (UID: \"db193bcc-748f-4830-873c-48a2f9229b40\") " pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:27:44 crc kubenswrapper[4813]: E0317 09:27:44.857224 4813 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 17 09:27:44 crc kubenswrapper[4813]: E0317 09:27:44.857269 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-webhook-certs podName:db193bcc-748f-4830-873c-48a2f9229b40 nodeName:}" failed. No retries permitted until 2026-03-17 09:27:45.35724764 +0000 UTC m=+1087.458051139 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-webhook-certs") pod "openstack-operator-controller-manager-65f8859ffd-mxfzh" (UID: "db193bcc-748f-4830-873c-48a2f9229b40") : secret "webhook-server-cert" not found Mar 17 09:27:44 crc kubenswrapper[4813]: E0317 09:27:44.857320 4813 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 17 09:27:44 crc kubenswrapper[4813]: E0317 09:27:44.857403 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-metrics-certs podName:db193bcc-748f-4830-873c-48a2f9229b40 nodeName:}" failed. No retries permitted until 2026-03-17 09:27:45.357381104 +0000 UTC m=+1087.458184703 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-metrics-certs") pod "openstack-operator-controller-manager-65f8859ffd-mxfzh" (UID: "db193bcc-748f-4830-873c-48a2f9229b40") : secret "metrics-server-cert" not found Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.880444 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phmjt\" (UniqueName: \"kubernetes.io/projected/db193bcc-748f-4830-873c-48a2f9229b40-kube-api-access-phmjt\") pod \"openstack-operator-controller-manager-65f8859ffd-mxfzh\" (UID: \"db193bcc-748f-4830-873c-48a2f9229b40\") " pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.882712 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bkzld\" (UniqueName: \"kubernetes.io/projected/bcd2954f-b240-4cc3-8387-15b0e64a9721-kube-api-access-bkzld\") pod \"rabbitmq-cluster-operator-manager-668c99d594-rxzjs\" (UID: \"bcd2954f-b240-4cc3-8387-15b0e64a9721\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rxzjs" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.910215 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-wrtl9" Mar 17 09:27:44 crc kubenswrapper[4813]: I0317 09:27:44.939531 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rxzjs" Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.061236 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8e80dcf7-dae9-4faa-ab3f-b547b570926b-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-g5vgd\" (UID: \"8e80dcf7-dae9-4faa-ab3f-b547b570926b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.061759 4813 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.061815 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e80dcf7-dae9-4faa-ab3f-b547b570926b-cert podName:8e80dcf7-dae9-4faa-ab3f-b547b570926b nodeName:}" failed. No retries permitted until 2026-03-17 09:27:46.061799055 +0000 UTC m=+1088.162602554 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8e80dcf7-dae9-4faa-ab3f-b547b570926b-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" (UID: "8e80dcf7-dae9-4faa-ab3f-b547b570926b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.146605 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-59bc569d95-xkfkm"] Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.169229 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-8d58dc466-9cqth"] Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.258357 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ngd7n"] Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.264971 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-768b96df4c-c966n"] Mar 17 09:27:45 crc kubenswrapper[4813]: W0317 09:27:45.267017 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0e3eba88_a4a4_4c6d_9e90_5dba41268494.slice/crio-1f188603f534fe320aba016cdb30c0b54fb2b01cd18d05bd85f0256980b1105c WatchSource:0}: Error finding container 1f188603f534fe320aba016cdb30c0b54fb2b01cd18d05bd85f0256980b1105c: Status 404 returned error can't find the container with id 1f188603f534fe320aba016cdb30c0b54fb2b01cd18d05bd85f0256980b1105c Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.365064 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-metrics-certs\") pod \"openstack-operator-controller-manager-65f8859ffd-mxfzh\" (UID: \"db193bcc-748f-4830-873c-48a2f9229b40\") " pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.365464 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-webhook-certs\") pod \"openstack-operator-controller-manager-65f8859ffd-mxfzh\" (UID: \"db193bcc-748f-4830-873c-48a2f9229b40\") " pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.365219 4813 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.365590 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-metrics-certs podName:db193bcc-748f-4830-873c-48a2f9229b40 nodeName:}" failed. No retries permitted until 2026-03-17 09:27:46.365572806 +0000 UTC m=+1088.466376315 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-metrics-certs") pod "openstack-operator-controller-manager-65f8859ffd-mxfzh" (UID: "db193bcc-748f-4830-873c-48a2f9229b40") : secret "metrics-server-cert" not found Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.365515 4813 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.365639 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-webhook-certs podName:db193bcc-748f-4830-873c-48a2f9229b40 nodeName:}" failed. No retries permitted until 2026-03-17 09:27:46.365630508 +0000 UTC m=+1088.466434007 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-webhook-certs") pod "openstack-operator-controller-manager-65f8859ffd-mxfzh" (UID: "db193bcc-748f-4830-873c-48a2f9229b40") : secret "webhook-server-cert" not found Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.464748 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6f787dddc9-4ds8c"] Mar 17 09:27:45 crc kubenswrapper[4813]: W0317 09:27:45.469148 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod80f14d5d_c8b9_4068_9469_ea8eff96e9e5.slice/crio-e22e60a5d80b3343a5eafa87ce34e948973c919855577a758239f645c54a1cfd WatchSource:0}: Error finding container e22e60a5d80b3343a5eafa87ce34e948973c919855577a758239f645c54a1cfd: Status 404 returned error can't find the container with id e22e60a5d80b3343a5eafa87ce34e948973c919855577a758239f645c54a1cfd Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.474946 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-67dd5f86f5-hng2s"] Mar 17 09:27:45 crc kubenswrapper[4813]: W0317 09:27:45.488720 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3c5a1075_8105_401b_afb0_00014dafed0a.slice/crio-3cf95fc84b29a36aebf8d6ad351726190f20bf2dd1ed105f8f733fa77da7750c WatchSource:0}: Error finding container 3cf95fc84b29a36aebf8d6ad351726190f20bf2dd1ed105f8f733fa77da7750c: Status 404 returned error can't find the container with id 3cf95fc84b29a36aebf8d6ad351726190f20bf2dd1ed105f8f733fa77da7750c Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.498337 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-79df6bcc97-wzmxx"] Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.510770 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-5b9f45d989-mps6t"] Mar 17 09:27:45 crc kubenswrapper[4813]: W0317 09:27:45.513265 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc6e1a616_6d48_4ef1_93cb_62caed2de963.slice/crio-00adce692058eafd1c7f1e372e0f3da55806b76f916a157672bed43318928110 WatchSource:0}: Error finding container 00adce692058eafd1c7f1e372e0f3da55806b76f916a157672bed43318928110: Status 404 returned error can't find the container with id 00adce692058eafd1c7f1e372e0f3da55806b76f916a157672bed43318928110 Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.554437 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:703ad3a2b749bce100f1e2a445312b65dc3b8b45e8c8ba59f311d3f8f3368113,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-d6868,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-8464cc45fb-tfhsc_openstack-operators(668691cc-39f3-4b04-8ed6-bc86f9965f12): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.555769 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-588d4d986b-tn4pv"] Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.555829 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-tfhsc" podUID="668691cc-39f3-4b04-8ed6-bc86f9965f12" Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.564711 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-767865f676-mnsb4"] Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.572755 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-8464cc45fb-tfhsc"] Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.583414 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-55f864c847-dbs68"] Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.629059 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-884679f54-7d66d"] Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.639475 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-c674c5965-njxx7"] Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.642017 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zv8d6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5c5cb9c4d7-r2ksk_openstack-operators(860ccadc-ca76-4523-9543-2b0960eb77a9): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.642388 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:866844c5b88e1e0518ceb7490cac9d093da3fb8b2f27ba7bd9bd89f946b9ee6e,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zthsr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-c674c5965-njxx7_openstack-operators(ac052585-f733-47d5-91b9-87ad7957e511): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.643251 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-r2ksk" podUID="860ccadc-ca76-4523-9543-2b0960eb77a9" Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.643745 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-c674c5965-njxx7" podUID="ac052585-f733-47d5-91b9-87ad7957e511" Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.649643 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5c5cb9c4d7-r2ksk"] Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.653640 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jnfhn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5784578c99-hcsvd_openstack-operators(968158c3-2dfc-439e-9d20-8706d611fec1): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.654819 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-hcsvd" podUID="968158c3-2dfc-439e-9d20-8706d611fec1" Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.674771 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c9ea4c59-d87b-4e58-ac8f-2786f8f32f02-cert\") pod \"infra-operator-controller-manager-7b9c774f96-7965b\" (UID: \"c9ea4c59-d87b-4e58-ac8f-2786f8f32f02\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-7965b" Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.676078 4813 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.676152 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9ea4c59-d87b-4e58-ac8f-2786f8f32f02-cert podName:c9ea4c59-d87b-4e58-ac8f-2786f8f32f02 nodeName:}" failed. No retries permitted until 2026-03-17 09:27:47.67613428 +0000 UTC m=+1089.776937779 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c9ea4c59-d87b-4e58-ac8f-2786f8f32f02-cert") pod "infra-operator-controller-manager-7b9c774f96-7965b" (UID: "c9ea4c59-d87b-4e58-ac8f-2786f8f32f02") : secret "infra-operator-webhook-server-cert" not found Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.678777 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-wrtl9"] Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.690045 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-5784578c99-hcsvd"] Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.690241 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-d6b694c5-52xx2"] Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.697766 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-5d488d59fb-tqg57"] Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.702889 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rxzjs"] Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.713886 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bfqtl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-5d488d59fb-tqg57_openstack-operators(b4b9a61e-819f-4da5-a28a-8e49c0a3f7cd): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.713934 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-b6wxb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-d6b694c5-52xx2_openstack-operators(ec48dbf8-730b-4a2e-9986-03d8f66de013): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.714111 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-xkjbm,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-6c4d75f7f9-wrtl9_openstack-operators(520d3ab3-78d7-4be3-8466-9dadfce15c2c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.714655 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bkzld,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-rxzjs_openstack-operators(bcd2954f-b240-4cc3-8387-15b0e64a9721): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.714975 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tqg57" podUID="b4b9a61e-819f-4da5-a28a-8e49c0a3f7cd" Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.715001 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-52xx2" podUID="ec48dbf8-730b-4a2e-9986-03d8f66de013" Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.715717 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-wrtl9" podUID="520d3ab3-78d7-4be3-8466-9dadfce15c2c" Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.715761 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rxzjs" podUID="bcd2954f-b240-4cc3-8387-15b0e64a9721" Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.912999 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-xkfkm" event={"ID":"227f29cf-ec7d-40e9-b28c-c1ccae9d880e","Type":"ContainerStarted","Data":"5323372b0c8236470c4f26800c60222ef9b68a61c66042a9873b7d0d27693f18"} Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.914081 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c966n" event={"ID":"fd3396be-4825-4edc-a41f-443af1413e3b","Type":"ContainerStarted","Data":"1271646e4bb71ef0562dbbbd45c8ec981777398fdb177934dade8ef6bb2efc71"} Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.915279 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-9cqth" event={"ID":"1b2b8088-31e5-49af-91b4-a3fbfe797af5","Type":"ContainerStarted","Data":"9653dd3276fda01c9b4ce4dfa62c94f3c8d65b0a616a10f057e519c223ec3cb2"} Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.916198 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-wzmxx" event={"ID":"3c5a1075-8105-401b-afb0-00014dafed0a","Type":"ContainerStarted","Data":"3cf95fc84b29a36aebf8d6ad351726190f20bf2dd1ed105f8f733fa77da7750c"} Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.917706 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tqg57" event={"ID":"b4b9a61e-819f-4da5-a28a-8e49c0a3f7cd","Type":"ContainerStarted","Data":"90dcbe2f0f2340fd2389c6ee79e4035f389a30ad8664a38d5f2af75c2a01da58"} Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.918487 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-tn4pv" event={"ID":"9b3cf477-8c47-4a38-beee-8bd3f70977e0","Type":"ContainerStarted","Data":"7cf01e618cb68f748dd8a4a3e417d1eaa8cd5a8692c25f84f84190fd94552301"} Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.919274 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a\\\"\"" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tqg57" podUID="b4b9a61e-819f-4da5-a28a-8e49c0a3f7cd" Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.920259 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-884679f54-7d66d" event={"ID":"76eb47fe-8f18-4f98-a8ba-d89169d0de93","Type":"ContainerStarted","Data":"e9c3d0cc1253365348481595a5b0eb7e1839811e510e683ebb3bd9f0bc5a8634"} Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.921338 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rxzjs" event={"ID":"bcd2954f-b240-4cc3-8387-15b0e64a9721","Type":"ContainerStarted","Data":"18ebe9fa1747f99d306424b27952dcdc4845013588fd73c99a0e41791289d790"} Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.922384 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rxzjs" podUID="bcd2954f-b240-4cc3-8387-15b0e64a9721" Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.922833 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ngd7n" event={"ID":"0e3eba88-a4a4-4c6d-9e90-5dba41268494","Type":"ContainerStarted","Data":"1f188603f534fe320aba016cdb30c0b54fb2b01cd18d05bd85f0256980b1105c"} Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.923943 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-mps6t" event={"ID":"c6e1a616-6d48-4ef1-93cb-62caed2de963","Type":"ContainerStarted","Data":"00adce692058eafd1c7f1e372e0f3da55806b76f916a157672bed43318928110"} Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.929301 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-hng2s" event={"ID":"de6c9679-9f9f-481d-85d9-218c9199088e","Type":"ContainerStarted","Data":"c0069980bb03eb74c3cf0a176a729a2c90a3dc3a45d6f62fd6ae31e48dfc405c"} Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.937952 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-wrtl9" event={"ID":"520d3ab3-78d7-4be3-8466-9dadfce15c2c","Type":"ContainerStarted","Data":"17a18698419ec905fb7708f2184cdd2c6ba2a02441cd4cd9169da8180746b5de"} Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.940108 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-wrtl9" podUID="520d3ab3-78d7-4be3-8466-9dadfce15c2c" Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.940914 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-r2ksk" event={"ID":"860ccadc-ca76-4523-9543-2b0960eb77a9","Type":"ContainerStarted","Data":"e9fd3928dea21d940f8df517895fa9e0ad3f15c3f17e63b2e477613e0218173d"} Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.942258 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42\\\"\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-r2ksk" podUID="860ccadc-ca76-4523-9543-2b0960eb77a9" Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.943506 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-c674c5965-njxx7" event={"ID":"ac052585-f733-47d5-91b9-87ad7957e511","Type":"ContainerStarted","Data":"3b194b5761dfdcb63e76315f382f7bea242d1a707c1022363f0089daa4a3ecf8"} Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.945136 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-tfhsc" event={"ID":"668691cc-39f3-4b04-8ed6-bc86f9965f12","Type":"ContainerStarted","Data":"08388a6c3227d70e248913a1e8ea21b2b6f73668b82b76997dbd5f968fb08786"} Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.945410 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:866844c5b88e1e0518ceb7490cac9d093da3fb8b2f27ba7bd9bd89f946b9ee6e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-c674c5965-njxx7" podUID="ac052585-f733-47d5-91b9-87ad7957e511" Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.946541 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-52xx2" event={"ID":"ec48dbf8-730b-4a2e-9986-03d8f66de013","Type":"ContainerStarted","Data":"8bec2e55d2f6709fa959417bc92064a08e7d10296096cde36bd28bc87fdc295e"} Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.947425 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:703ad3a2b749bce100f1e2a445312b65dc3b8b45e8c8ba59f311d3f8f3368113\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-tfhsc" podUID="668691cc-39f3-4b04-8ed6-bc86f9965f12" Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.950518 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-52xx2" podUID="ec48dbf8-730b-4a2e-9986-03d8f66de013" Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.951260 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5784578c99-hcsvd" event={"ID":"968158c3-2dfc-439e-9d20-8706d611fec1","Type":"ContainerStarted","Data":"ff06b198d7a6f8614efad48b05952e843cffc78a6409b8b99d2e83a2effa2167"} Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.953249 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-4ds8c" event={"ID":"80f14d5d-c8b9-4068-9469-ea8eff96e9e5","Type":"ContainerStarted","Data":"e22e60a5d80b3343a5eafa87ce34e948973c919855577a758239f645c54a1cfd"} Mar 17 09:27:45 crc kubenswrapper[4813]: E0317 09:27:45.958124 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-hcsvd" podUID="968158c3-2dfc-439e-9d20-8706d611fec1" Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.960324 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-55f864c847-dbs68" event={"ID":"5e7b137d-9b58-448b-91a4-3a069dfb4d10","Type":"ContainerStarted","Data":"68966ba4f5132c964c05a9e6cc0e3b6a5f8c54052d270e2ef65535b06ed81a32"} Mar 17 09:27:45 crc kubenswrapper[4813]: I0317 09:27:45.965849 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-767865f676-mnsb4" event={"ID":"c3419efc-507c-46bc-aedb-ef5eb3f0f917","Type":"ContainerStarted","Data":"6e01c20054dbdb88b93b75b8dcf9cfc5da8a1fdaab9849a1cff26fcadc49df99"} Mar 17 09:27:46 crc kubenswrapper[4813]: I0317 09:27:46.081683 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8e80dcf7-dae9-4faa-ab3f-b547b570926b-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-g5vgd\" (UID: \"8e80dcf7-dae9-4faa-ab3f-b547b570926b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" Mar 17 09:27:46 crc kubenswrapper[4813]: E0317 09:27:46.081812 4813 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 09:27:46 crc kubenswrapper[4813]: E0317 09:27:46.081860 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e80dcf7-dae9-4faa-ab3f-b547b570926b-cert podName:8e80dcf7-dae9-4faa-ab3f-b547b570926b nodeName:}" failed. No retries permitted until 2026-03-17 09:27:48.081845753 +0000 UTC m=+1090.182649252 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8e80dcf7-dae9-4faa-ab3f-b547b570926b-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" (UID: "8e80dcf7-dae9-4faa-ab3f-b547b570926b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 09:27:46 crc kubenswrapper[4813]: I0317 09:27:46.401185 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-webhook-certs\") pod \"openstack-operator-controller-manager-65f8859ffd-mxfzh\" (UID: \"db193bcc-748f-4830-873c-48a2f9229b40\") " pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:27:46 crc kubenswrapper[4813]: I0317 09:27:46.401534 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-metrics-certs\") pod \"openstack-operator-controller-manager-65f8859ffd-mxfzh\" (UID: \"db193bcc-748f-4830-873c-48a2f9229b40\") " pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:27:46 crc kubenswrapper[4813]: E0317 09:27:46.401368 4813 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 17 09:27:46 crc kubenswrapper[4813]: E0317 09:27:46.401662 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-webhook-certs podName:db193bcc-748f-4830-873c-48a2f9229b40 nodeName:}" failed. No retries permitted until 2026-03-17 09:27:48.401639337 +0000 UTC m=+1090.502442906 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-webhook-certs") pod "openstack-operator-controller-manager-65f8859ffd-mxfzh" (UID: "db193bcc-748f-4830-873c-48a2f9229b40") : secret "webhook-server-cert" not found Mar 17 09:27:46 crc kubenswrapper[4813]: E0317 09:27:46.401723 4813 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 17 09:27:46 crc kubenswrapper[4813]: E0317 09:27:46.401775 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-metrics-certs podName:db193bcc-748f-4830-873c-48a2f9229b40 nodeName:}" failed. No retries permitted until 2026-03-17 09:27:48.401760321 +0000 UTC m=+1090.502563820 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-metrics-certs") pod "openstack-operator-controller-manager-65f8859ffd-mxfzh" (UID: "db193bcc-748f-4830-873c-48a2f9229b40") : secret "metrics-server-cert" not found Mar 17 09:27:46 crc kubenswrapper[4813]: E0317 09:27:46.973335 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:d9c55e8c6304a0e32289b5e8c69a87ea59b9968918a5c85b7c384633df82c807\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-wrtl9" podUID="520d3ab3-78d7-4be3-8466-9dadfce15c2c" Mar 17 09:27:46 crc kubenswrapper[4813]: E0317 09:27:46.973953 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/horizon-operator@sha256:703ad3a2b749bce100f1e2a445312b65dc3b8b45e8c8ba59f311d3f8f3368113\\\"\"" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-tfhsc" podUID="668691cc-39f3-4b04-8ed6-bc86f9965f12" Mar 17 09:27:46 crc kubenswrapper[4813]: E0317 09:27:46.974143 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:c500fa7080b94105e85eeced772d8872e4168904e74ba02116e15ab66f522444\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-52xx2" podUID="ec48dbf8-730b-4a2e-9986-03d8f66de013" Mar 17 09:27:46 crc kubenswrapper[4813]: E0317 09:27:46.974274 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rxzjs" podUID="bcd2954f-b240-4cc3-8387-15b0e64a9721" Mar 17 09:27:46 crc kubenswrapper[4813]: E0317 09:27:46.974498 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:7398eb8fa5a4844d3326a5dff759d17199870c389b3ce3011a038b27bf95512a\\\"\"" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tqg57" podUID="b4b9a61e-819f-4da5-a28a-8e49c0a3f7cd" Mar 17 09:27:46 crc kubenswrapper[4813]: E0317 09:27:46.974519 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42\\\"\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-r2ksk" podUID="860ccadc-ca76-4523-9543-2b0960eb77a9" Mar 17 09:27:46 crc kubenswrapper[4813]: E0317 09:27:46.975089 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:866844c5b88e1e0518ceb7490cac9d093da3fb8b2f27ba7bd9bd89f946b9ee6e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-c674c5965-njxx7" podUID="ac052585-f733-47d5-91b9-87ad7957e511" Mar 17 09:27:46 crc kubenswrapper[4813]: E0317 09:27:46.975257 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-hcsvd" podUID="968158c3-2dfc-439e-9d20-8706d611fec1" Mar 17 09:27:47 crc kubenswrapper[4813]: I0317 09:27:47.725380 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c9ea4c59-d87b-4e58-ac8f-2786f8f32f02-cert\") pod \"infra-operator-controller-manager-7b9c774f96-7965b\" (UID: \"c9ea4c59-d87b-4e58-ac8f-2786f8f32f02\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-7965b" Mar 17 09:27:47 crc kubenswrapper[4813]: E0317 09:27:47.725662 4813 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 17 09:27:47 crc kubenswrapper[4813]: E0317 09:27:47.725813 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9ea4c59-d87b-4e58-ac8f-2786f8f32f02-cert podName:c9ea4c59-d87b-4e58-ac8f-2786f8f32f02 nodeName:}" failed. No retries permitted until 2026-03-17 09:27:51.725785116 +0000 UTC m=+1093.826588615 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c9ea4c59-d87b-4e58-ac8f-2786f8f32f02-cert") pod "infra-operator-controller-manager-7b9c774f96-7965b" (UID: "c9ea4c59-d87b-4e58-ac8f-2786f8f32f02") : secret "infra-operator-webhook-server-cert" not found Mar 17 09:27:48 crc kubenswrapper[4813]: I0317 09:27:48.130234 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8e80dcf7-dae9-4faa-ab3f-b547b570926b-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-g5vgd\" (UID: \"8e80dcf7-dae9-4faa-ab3f-b547b570926b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" Mar 17 09:27:48 crc kubenswrapper[4813]: E0317 09:27:48.131494 4813 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 09:27:48 crc kubenswrapper[4813]: E0317 09:27:48.131555 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e80dcf7-dae9-4faa-ab3f-b547b570926b-cert podName:8e80dcf7-dae9-4faa-ab3f-b547b570926b nodeName:}" failed. No retries permitted until 2026-03-17 09:27:52.13154143 +0000 UTC m=+1094.232344929 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8e80dcf7-dae9-4faa-ab3f-b547b570926b-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" (UID: "8e80dcf7-dae9-4faa-ab3f-b547b570926b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 09:27:48 crc kubenswrapper[4813]: I0317 09:27:48.434404 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-webhook-certs\") pod \"openstack-operator-controller-manager-65f8859ffd-mxfzh\" (UID: \"db193bcc-748f-4830-873c-48a2f9229b40\") " pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:27:48 crc kubenswrapper[4813]: I0317 09:27:48.434839 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-metrics-certs\") pod \"openstack-operator-controller-manager-65f8859ffd-mxfzh\" (UID: \"db193bcc-748f-4830-873c-48a2f9229b40\") " pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:27:48 crc kubenswrapper[4813]: E0317 09:27:48.434683 4813 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 17 09:27:48 crc kubenswrapper[4813]: E0317 09:27:48.435111 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-webhook-certs podName:db193bcc-748f-4830-873c-48a2f9229b40 nodeName:}" failed. No retries permitted until 2026-03-17 09:27:52.435093963 +0000 UTC m=+1094.535897472 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-webhook-certs") pod "openstack-operator-controller-manager-65f8859ffd-mxfzh" (UID: "db193bcc-748f-4830-873c-48a2f9229b40") : secret "webhook-server-cert" not found Mar 17 09:27:48 crc kubenswrapper[4813]: E0317 09:27:48.435051 4813 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 17 09:27:48 crc kubenswrapper[4813]: E0317 09:27:48.435483 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-metrics-certs podName:db193bcc-748f-4830-873c-48a2f9229b40 nodeName:}" failed. No retries permitted until 2026-03-17 09:27:52.435471655 +0000 UTC m=+1094.536275164 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-metrics-certs") pod "openstack-operator-controller-manager-65f8859ffd-mxfzh" (UID: "db193bcc-748f-4830-873c-48a2f9229b40") : secret "metrics-server-cert" not found Mar 17 09:27:48 crc kubenswrapper[4813]: I0317 09:27:48.980887 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-sql9s"] Mar 17 09:27:48 crc kubenswrapper[4813]: I0317 09:27:48.984305 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sql9s" Mar 17 09:27:48 crc kubenswrapper[4813]: I0317 09:27:48.995381 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sql9s"] Mar 17 09:27:49 crc kubenswrapper[4813]: I0317 09:27:49.145043 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xnqb\" (UniqueName: \"kubernetes.io/projected/19be6482-9c7f-465b-8a3d-5b99669fa87d-kube-api-access-8xnqb\") pod \"redhat-marketplace-sql9s\" (UID: \"19be6482-9c7f-465b-8a3d-5b99669fa87d\") " pod="openshift-marketplace/redhat-marketplace-sql9s" Mar 17 09:27:49 crc kubenswrapper[4813]: I0317 09:27:49.145179 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19be6482-9c7f-465b-8a3d-5b99669fa87d-catalog-content\") pod \"redhat-marketplace-sql9s\" (UID: \"19be6482-9c7f-465b-8a3d-5b99669fa87d\") " pod="openshift-marketplace/redhat-marketplace-sql9s" Mar 17 09:27:49 crc kubenswrapper[4813]: I0317 09:27:49.145267 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19be6482-9c7f-465b-8a3d-5b99669fa87d-utilities\") pod \"redhat-marketplace-sql9s\" (UID: \"19be6482-9c7f-465b-8a3d-5b99669fa87d\") " pod="openshift-marketplace/redhat-marketplace-sql9s" Mar 17 09:27:49 crc kubenswrapper[4813]: I0317 09:27:49.246877 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xnqb\" (UniqueName: \"kubernetes.io/projected/19be6482-9c7f-465b-8a3d-5b99669fa87d-kube-api-access-8xnqb\") pod \"redhat-marketplace-sql9s\" (UID: \"19be6482-9c7f-465b-8a3d-5b99669fa87d\") " pod="openshift-marketplace/redhat-marketplace-sql9s" Mar 17 09:27:49 crc kubenswrapper[4813]: I0317 09:27:49.246966 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19be6482-9c7f-465b-8a3d-5b99669fa87d-catalog-content\") pod \"redhat-marketplace-sql9s\" (UID: \"19be6482-9c7f-465b-8a3d-5b99669fa87d\") " pod="openshift-marketplace/redhat-marketplace-sql9s" Mar 17 09:27:49 crc kubenswrapper[4813]: I0317 09:27:49.247046 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19be6482-9c7f-465b-8a3d-5b99669fa87d-utilities\") pod \"redhat-marketplace-sql9s\" (UID: \"19be6482-9c7f-465b-8a3d-5b99669fa87d\") " pod="openshift-marketplace/redhat-marketplace-sql9s" Mar 17 09:27:49 crc kubenswrapper[4813]: I0317 09:27:49.247830 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19be6482-9c7f-465b-8a3d-5b99669fa87d-utilities\") pod \"redhat-marketplace-sql9s\" (UID: \"19be6482-9c7f-465b-8a3d-5b99669fa87d\") " pod="openshift-marketplace/redhat-marketplace-sql9s" Mar 17 09:27:49 crc kubenswrapper[4813]: I0317 09:27:49.247931 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19be6482-9c7f-465b-8a3d-5b99669fa87d-catalog-content\") pod \"redhat-marketplace-sql9s\" (UID: \"19be6482-9c7f-465b-8a3d-5b99669fa87d\") " pod="openshift-marketplace/redhat-marketplace-sql9s" Mar 17 09:27:49 crc kubenswrapper[4813]: I0317 09:27:49.268281 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xnqb\" (UniqueName: \"kubernetes.io/projected/19be6482-9c7f-465b-8a3d-5b99669fa87d-kube-api-access-8xnqb\") pod \"redhat-marketplace-sql9s\" (UID: \"19be6482-9c7f-465b-8a3d-5b99669fa87d\") " pod="openshift-marketplace/redhat-marketplace-sql9s" Mar 17 09:27:49 crc kubenswrapper[4813]: I0317 09:27:49.315072 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sql9s" Mar 17 09:27:51 crc kubenswrapper[4813]: I0317 09:27:51.783404 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c9ea4c59-d87b-4e58-ac8f-2786f8f32f02-cert\") pod \"infra-operator-controller-manager-7b9c774f96-7965b\" (UID: \"c9ea4c59-d87b-4e58-ac8f-2786f8f32f02\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-7965b" Mar 17 09:27:51 crc kubenswrapper[4813]: E0317 09:27:51.783638 4813 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Mar 17 09:27:51 crc kubenswrapper[4813]: E0317 09:27:51.783922 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c9ea4c59-d87b-4e58-ac8f-2786f8f32f02-cert podName:c9ea4c59-d87b-4e58-ac8f-2786f8f32f02 nodeName:}" failed. No retries permitted until 2026-03-17 09:27:59.783898983 +0000 UTC m=+1101.884702552 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/c9ea4c59-d87b-4e58-ac8f-2786f8f32f02-cert") pod "infra-operator-controller-manager-7b9c774f96-7965b" (UID: "c9ea4c59-d87b-4e58-ac8f-2786f8f32f02") : secret "infra-operator-webhook-server-cert" not found Mar 17 09:27:52 crc kubenswrapper[4813]: I0317 09:27:52.192310 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8e80dcf7-dae9-4faa-ab3f-b547b570926b-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-g5vgd\" (UID: \"8e80dcf7-dae9-4faa-ab3f-b547b570926b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" Mar 17 09:27:52 crc kubenswrapper[4813]: E0317 09:27:52.192506 4813 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 09:27:52 crc kubenswrapper[4813]: E0317 09:27:52.192701 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e80dcf7-dae9-4faa-ab3f-b547b570926b-cert podName:8e80dcf7-dae9-4faa-ab3f-b547b570926b nodeName:}" failed. No retries permitted until 2026-03-17 09:28:00.192677362 +0000 UTC m=+1102.293480871 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8e80dcf7-dae9-4faa-ab3f-b547b570926b-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" (UID: "8e80dcf7-dae9-4faa-ab3f-b547b570926b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 09:27:52 crc kubenswrapper[4813]: I0317 09:27:52.497065 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-webhook-certs\") pod \"openstack-operator-controller-manager-65f8859ffd-mxfzh\" (UID: \"db193bcc-748f-4830-873c-48a2f9229b40\") " pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:27:52 crc kubenswrapper[4813]: I0317 09:27:52.497198 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-metrics-certs\") pod \"openstack-operator-controller-manager-65f8859ffd-mxfzh\" (UID: \"db193bcc-748f-4830-873c-48a2f9229b40\") " pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:27:52 crc kubenswrapper[4813]: E0317 09:27:52.497254 4813 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 17 09:27:52 crc kubenswrapper[4813]: E0317 09:27:52.497327 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-webhook-certs podName:db193bcc-748f-4830-873c-48a2f9229b40 nodeName:}" failed. No retries permitted until 2026-03-17 09:28:00.497309459 +0000 UTC m=+1102.598112958 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-webhook-certs") pod "openstack-operator-controller-manager-65f8859ffd-mxfzh" (UID: "db193bcc-748f-4830-873c-48a2f9229b40") : secret "webhook-server-cert" not found Mar 17 09:27:52 crc kubenswrapper[4813]: E0317 09:27:52.497346 4813 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 17 09:27:52 crc kubenswrapper[4813]: E0317 09:27:52.497408 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-metrics-certs podName:db193bcc-748f-4830-873c-48a2f9229b40 nodeName:}" failed. No retries permitted until 2026-03-17 09:28:00.497389552 +0000 UTC m=+1102.598193151 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-metrics-certs") pod "openstack-operator-controller-manager-65f8859ffd-mxfzh" (UID: "db193bcc-748f-4830-873c-48a2f9229b40") : secret "metrics-server-cert" not found Mar 17 09:27:57 crc kubenswrapper[4813]: I0317 09:27:57.676504 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-sql9s"] Mar 17 09:27:57 crc kubenswrapper[4813]: W0317 09:27:57.687690 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod19be6482_9c7f_465b_8a3d_5b99669fa87d.slice/crio-15ea27d29d16b3f457ec3fb7c97da716ced37ad5022a0378ed44d37c2167af51 WatchSource:0}: Error finding container 15ea27d29d16b3f457ec3fb7c97da716ced37ad5022a0378ed44d37c2167af51: Status 404 returned error can't find the container with id 15ea27d29d16b3f457ec3fb7c97da716ced37ad5022a0378ed44d37c2167af51 Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.062680 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-4ds8c" event={"ID":"80f14d5d-c8b9-4068-9469-ea8eff96e9e5","Type":"ContainerStarted","Data":"6902a7025d01635e0ee345220e76ef3517320942ff18bd1aa8e72c449444a042"} Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.063181 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-4ds8c" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.072042 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-55f864c847-dbs68" event={"ID":"5e7b137d-9b58-448b-91a4-3a069dfb4d10","Type":"ContainerStarted","Data":"77c6a06ad3fe2714a7e5eaadfc571cd6f007920c35c2e6431e716758b1302c63"} Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.072578 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-55f864c847-dbs68" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.074753 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-tn4pv" event={"ID":"9b3cf477-8c47-4a38-beee-8bd3f70977e0","Type":"ContainerStarted","Data":"46b92ebfc2d4a23391f458695ff2efef3183235cb1fb81b2f92b7cfc1c56ad38"} Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.075373 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-tn4pv" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.087813 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c966n" event={"ID":"fd3396be-4825-4edc-a41f-443af1413e3b","Type":"ContainerStarted","Data":"498fc00dfeed0fc781738442d58cc530919b6554ae0749536695d88250e11117"} Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.087888 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c966n" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.088807 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-4ds8c" podStartSLOduration=3.271571131 podStartE2EDuration="15.088790777s" podCreationTimestamp="2026-03-17 09:27:43 +0000 UTC" firstStartedPulling="2026-03-17 09:27:45.471132481 +0000 UTC m=+1087.571935980" lastFinishedPulling="2026-03-17 09:27:57.288352127 +0000 UTC m=+1099.389155626" observedRunningTime="2026-03-17 09:27:58.08409013 +0000 UTC m=+1100.184893629" watchObservedRunningTime="2026-03-17 09:27:58.088790777 +0000 UTC m=+1100.189594276" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.103320 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-9cqth" event={"ID":"1b2b8088-31e5-49af-91b4-a3fbfe797af5","Type":"ContainerStarted","Data":"f999403a1846541b3c1169d56b2576de8010849f0fd8494084373076f0febc3b"} Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.103963 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-9cqth" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.111932 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-55f864c847-dbs68" podStartSLOduration=3.4192412 podStartE2EDuration="15.111917644s" podCreationTimestamp="2026-03-17 09:27:43 +0000 UTC" firstStartedPulling="2026-03-17 09:27:45.552062223 +0000 UTC m=+1087.652865722" lastFinishedPulling="2026-03-17 09:27:57.244738667 +0000 UTC m=+1099.345542166" observedRunningTime="2026-03-17 09:27:58.111097518 +0000 UTC m=+1100.211901017" watchObservedRunningTime="2026-03-17 09:27:58.111917644 +0000 UTC m=+1100.212721133" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.123265 4813 generic.go:334] "Generic (PLEG): container finished" podID="19be6482-9c7f-465b-8a3d-5b99669fa87d" containerID="68eeb55595f33e79510e97fd8c2eb0064317b08a9de1a31c9069261af29e521a" exitCode=0 Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.123323 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sql9s" event={"ID":"19be6482-9c7f-465b-8a3d-5b99669fa87d","Type":"ContainerDied","Data":"68eeb55595f33e79510e97fd8c2eb0064317b08a9de1a31c9069261af29e521a"} Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.123346 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sql9s" event={"ID":"19be6482-9c7f-465b-8a3d-5b99669fa87d","Type":"ContainerStarted","Data":"15ea27d29d16b3f457ec3fb7c97da716ced37ad5022a0378ed44d37c2167af51"} Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.138528 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-884679f54-7d66d" event={"ID":"76eb47fe-8f18-4f98-a8ba-d89169d0de93","Type":"ContainerStarted","Data":"dfec4594c042e55ce76ad0dcf7644742147263177efe9b364eb290efae122fc6"} Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.139179 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-884679f54-7d66d" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.147364 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-tn4pv" podStartSLOduration=3.39954401 podStartE2EDuration="15.147349176s" podCreationTimestamp="2026-03-17 09:27:43 +0000 UTC" firstStartedPulling="2026-03-17 09:27:45.543120992 +0000 UTC m=+1087.643924491" lastFinishedPulling="2026-03-17 09:27:57.290926158 +0000 UTC m=+1099.391729657" observedRunningTime="2026-03-17 09:27:58.145015423 +0000 UTC m=+1100.245818922" watchObservedRunningTime="2026-03-17 09:27:58.147349176 +0000 UTC m=+1100.248152665" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.155815 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ngd7n" event={"ID":"0e3eba88-a4a4-4c6d-9e90-5dba41268494","Type":"ContainerStarted","Data":"792f187373a17f2b77105d6118a3d98e5ed3a9fddc61eb6ad2eb1704c125daae"} Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.156448 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ngd7n" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.169465 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-mps6t" event={"ID":"c6e1a616-6d48-4ef1-93cb-62caed2de963","Type":"ContainerStarted","Data":"f7c957b8addf60bb864d751647e57ab85b3c6f6cee5a1bc5c9071a9894c68c31"} Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.169517 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-mps6t" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.172245 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-9cqth" podStartSLOduration=3.096613586 podStartE2EDuration="15.172227947s" podCreationTimestamp="2026-03-17 09:27:43 +0000 UTC" firstStartedPulling="2026-03-17 09:27:45.188159134 +0000 UTC m=+1087.288962633" lastFinishedPulling="2026-03-17 09:27:57.263773495 +0000 UTC m=+1099.364576994" observedRunningTime="2026-03-17 09:27:58.169733599 +0000 UTC m=+1100.270537098" watchObservedRunningTime="2026-03-17 09:27:58.172227947 +0000 UTC m=+1100.273031446" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.190539 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-xkfkm" event={"ID":"227f29cf-ec7d-40e9-b28c-c1ccae9d880e","Type":"ContainerStarted","Data":"546499b16eb0062217ca09331cbe4f4f60322bfbcb669468f5817f99177aeedf"} Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.192151 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-xkfkm" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.216555 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-hng2s" event={"ID":"de6c9679-9f9f-481d-85d9-218c9199088e","Type":"ContainerStarted","Data":"511e910f01d2a7ed6be8b286b5dc72c097d81c6c01153cd5bd474340dc219c8a"} Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.217135 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-hng2s" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.227581 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c966n" podStartSLOduration=3.160026048 podStartE2EDuration="15.227562705s" podCreationTimestamp="2026-03-17 09:27:43 +0000 UTC" firstStartedPulling="2026-03-17 09:27:45.280875516 +0000 UTC m=+1087.381679015" lastFinishedPulling="2026-03-17 09:27:57.348412173 +0000 UTC m=+1099.449215672" observedRunningTime="2026-03-17 09:27:58.224131008 +0000 UTC m=+1100.324934507" watchObservedRunningTime="2026-03-17 09:27:58.227562705 +0000 UTC m=+1100.328366204" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.230988 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-wzmxx" event={"ID":"3c5a1075-8105-401b-afb0-00014dafed0a","Type":"ContainerStarted","Data":"3660f17c1261d2154df633ff2c28342aac8abc21c0c27395c9ed91128cdfcab5"} Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.231678 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-wzmxx" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.244766 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-767865f676-mnsb4" event={"ID":"c3419efc-507c-46bc-aedb-ef5eb3f0f917","Type":"ContainerStarted","Data":"5ca1fb4c44a097a1f9f810d8458c9c9fb55e98a12210b21a0adcaccde91af5d0"} Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.245409 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-767865f676-mnsb4" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.259507 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-884679f54-7d66d" podStartSLOduration=2.611641262 podStartE2EDuration="14.259484138s" podCreationTimestamp="2026-03-17 09:27:44 +0000 UTC" firstStartedPulling="2026-03-17 09:27:45.64174662 +0000 UTC m=+1087.742550119" lastFinishedPulling="2026-03-17 09:27:57.289589496 +0000 UTC m=+1099.390392995" observedRunningTime="2026-03-17 09:27:58.255487962 +0000 UTC m=+1100.356291461" watchObservedRunningTime="2026-03-17 09:27:58.259484138 +0000 UTC m=+1100.360287647" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.289853 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-hng2s" podStartSLOduration=3.502415942 podStartE2EDuration="15.289836851s" podCreationTimestamp="2026-03-17 09:27:43 +0000 UTC" firstStartedPulling="2026-03-17 09:27:45.502944281 +0000 UTC m=+1087.603747780" lastFinishedPulling="2026-03-17 09:27:57.29036519 +0000 UTC m=+1099.391168689" observedRunningTime="2026-03-17 09:27:58.285823506 +0000 UTC m=+1100.386627005" watchObservedRunningTime="2026-03-17 09:27:58.289836851 +0000 UTC m=+1100.390640350" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.380797 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-xkfkm" podStartSLOduration=3.304478745 podStartE2EDuration="15.380779068s" podCreationTimestamp="2026-03-17 09:27:43 +0000 UTC" firstStartedPulling="2026-03-17 09:27:45.187823203 +0000 UTC m=+1087.288626702" lastFinishedPulling="2026-03-17 09:27:57.264123506 +0000 UTC m=+1099.364927025" observedRunningTime="2026-03-17 09:27:58.371542928 +0000 UTC m=+1100.472346427" watchObservedRunningTime="2026-03-17 09:27:58.380779068 +0000 UTC m=+1100.481582557" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.384297 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-wzmxx" podStartSLOduration=3.635891023 podStartE2EDuration="15.384286337s" podCreationTimestamp="2026-03-17 09:27:43 +0000 UTC" firstStartedPulling="2026-03-17 09:27:45.496362784 +0000 UTC m=+1087.597166303" lastFinishedPulling="2026-03-17 09:27:57.244758118 +0000 UTC m=+1099.345561617" observedRunningTime="2026-03-17 09:27:58.352449198 +0000 UTC m=+1100.453252697" watchObservedRunningTime="2026-03-17 09:27:58.384286337 +0000 UTC m=+1100.485089836" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.451361 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ngd7n" podStartSLOduration=3.430758501 podStartE2EDuration="15.451346705s" podCreationTimestamp="2026-03-17 09:27:43 +0000 UTC" firstStartedPulling="2026-03-17 09:27:45.268964061 +0000 UTC m=+1087.369767560" lastFinishedPulling="2026-03-17 09:27:57.289552255 +0000 UTC m=+1099.390355764" observedRunningTime="2026-03-17 09:27:58.435249288 +0000 UTC m=+1100.536052777" watchObservedRunningTime="2026-03-17 09:27:58.451346705 +0000 UTC m=+1100.552150204" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.451686 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-767865f676-mnsb4" podStartSLOduration=3.719967714 podStartE2EDuration="15.451681954s" podCreationTimestamp="2026-03-17 09:27:43 +0000 UTC" firstStartedPulling="2026-03-17 09:27:45.532476358 +0000 UTC m=+1087.633279857" lastFinishedPulling="2026-03-17 09:27:57.264190568 +0000 UTC m=+1099.364994097" observedRunningTime="2026-03-17 09:27:58.450632312 +0000 UTC m=+1100.551435801" watchObservedRunningTime="2026-03-17 09:27:58.451681954 +0000 UTC m=+1100.552485453" Mar 17 09:27:58 crc kubenswrapper[4813]: I0317 09:27:58.473202 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-mps6t" podStartSLOduration=2.7251985 podStartE2EDuration="14.47318287s" podCreationTimestamp="2026-03-17 09:27:44 +0000 UTC" firstStartedPulling="2026-03-17 09:27:45.51691305 +0000 UTC m=+1087.617716569" lastFinishedPulling="2026-03-17 09:27:57.26489744 +0000 UTC m=+1099.365700939" observedRunningTime="2026-03-17 09:27:58.471393833 +0000 UTC m=+1100.572197332" watchObservedRunningTime="2026-03-17 09:27:58.47318287 +0000 UTC m=+1100.573986369" Mar 17 09:27:59 crc kubenswrapper[4813]: I0317 09:27:59.252544 4813 generic.go:334] "Generic (PLEG): container finished" podID="19be6482-9c7f-465b-8a3d-5b99669fa87d" containerID="8762c2aa69313afeacf3cda72bbc3bfff53d87f94d3acbca115e5c7eae6abaf9" exitCode=0 Mar 17 09:27:59 crc kubenswrapper[4813]: I0317 09:27:59.252587 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sql9s" event={"ID":"19be6482-9c7f-465b-8a3d-5b99669fa87d","Type":"ContainerDied","Data":"8762c2aa69313afeacf3cda72bbc3bfff53d87f94d3acbca115e5c7eae6abaf9"} Mar 17 09:27:59 crc kubenswrapper[4813]: I0317 09:27:59.832437 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c9ea4c59-d87b-4e58-ac8f-2786f8f32f02-cert\") pod \"infra-operator-controller-manager-7b9c774f96-7965b\" (UID: \"c9ea4c59-d87b-4e58-ac8f-2786f8f32f02\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-7965b" Mar 17 09:27:59 crc kubenswrapper[4813]: I0317 09:27:59.847186 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/c9ea4c59-d87b-4e58-ac8f-2786f8f32f02-cert\") pod \"infra-operator-controller-manager-7b9c774f96-7965b\" (UID: \"c9ea4c59-d87b-4e58-ac8f-2786f8f32f02\") " pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-7965b" Mar 17 09:27:59 crc kubenswrapper[4813]: I0317 09:27:59.857752 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-7965b" Mar 17 09:28:00 crc kubenswrapper[4813]: I0317 09:28:00.132510 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562328-pmv66"] Mar 17 09:28:00 crc kubenswrapper[4813]: I0317 09:28:00.133848 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562328-pmv66" Mar 17 09:28:00 crc kubenswrapper[4813]: I0317 09:28:00.136616 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:28:00 crc kubenswrapper[4813]: I0317 09:28:00.137542 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:28:00 crc kubenswrapper[4813]: I0317 09:28:00.137698 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:28:00 crc kubenswrapper[4813]: I0317 09:28:00.143193 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562328-pmv66"] Mar 17 09:28:00 crc kubenswrapper[4813]: I0317 09:28:00.241159 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8e80dcf7-dae9-4faa-ab3f-b547b570926b-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-g5vgd\" (UID: \"8e80dcf7-dae9-4faa-ab3f-b547b570926b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" Mar 17 09:28:00 crc kubenswrapper[4813]: I0317 09:28:00.241340 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g27vx\" (UniqueName: \"kubernetes.io/projected/02f74478-75b1-4e8a-b347-d0bc8cc32f4c-kube-api-access-g27vx\") pod \"auto-csr-approver-29562328-pmv66\" (UID: \"02f74478-75b1-4e8a-b347-d0bc8cc32f4c\") " pod="openshift-infra/auto-csr-approver-29562328-pmv66" Mar 17 09:28:00 crc kubenswrapper[4813]: E0317 09:28:00.241360 4813 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 09:28:00 crc kubenswrapper[4813]: E0317 09:28:00.241435 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/8e80dcf7-dae9-4faa-ab3f-b547b570926b-cert podName:8e80dcf7-dae9-4faa-ab3f-b547b570926b nodeName:}" failed. No retries permitted until 2026-03-17 09:28:16.241413276 +0000 UTC m=+1118.342216765 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/8e80dcf7-dae9-4faa-ab3f-b547b570926b-cert") pod "openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" (UID: "8e80dcf7-dae9-4faa-ab3f-b547b570926b") : secret "openstack-baremetal-operator-webhook-server-cert" not found Mar 17 09:28:00 crc kubenswrapper[4813]: I0317 09:28:00.265499 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sql9s" event={"ID":"19be6482-9c7f-465b-8a3d-5b99669fa87d","Type":"ContainerStarted","Data":"2b0c94d67f86b72c7cd559d403934130305882ed950f70faf934bbcd8fe89097"} Mar 17 09:28:00 crc kubenswrapper[4813]: I0317 09:28:00.289612 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-sql9s" podStartSLOduration=10.474525352 podStartE2EDuration="12.289571959s" podCreationTimestamp="2026-03-17 09:27:48 +0000 UTC" firstStartedPulling="2026-03-17 09:27:58.128778783 +0000 UTC m=+1100.229582282" lastFinishedPulling="2026-03-17 09:27:59.94382539 +0000 UTC m=+1102.044628889" observedRunningTime="2026-03-17 09:28:00.284908873 +0000 UTC m=+1102.385712392" watchObservedRunningTime="2026-03-17 09:28:00.289571959 +0000 UTC m=+1102.390375468" Mar 17 09:28:00 crc kubenswrapper[4813]: I0317 09:28:00.309871 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-7b9c774f96-7965b"] Mar 17 09:28:00 crc kubenswrapper[4813]: I0317 09:28:00.345484 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g27vx\" (UniqueName: \"kubernetes.io/projected/02f74478-75b1-4e8a-b347-d0bc8cc32f4c-kube-api-access-g27vx\") pod \"auto-csr-approver-29562328-pmv66\" (UID: \"02f74478-75b1-4e8a-b347-d0bc8cc32f4c\") " pod="openshift-infra/auto-csr-approver-29562328-pmv66" Mar 17 09:28:00 crc kubenswrapper[4813]: I0317 09:28:00.380416 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g27vx\" (UniqueName: \"kubernetes.io/projected/02f74478-75b1-4e8a-b347-d0bc8cc32f4c-kube-api-access-g27vx\") pod \"auto-csr-approver-29562328-pmv66\" (UID: \"02f74478-75b1-4e8a-b347-d0bc8cc32f4c\") " pod="openshift-infra/auto-csr-approver-29562328-pmv66" Mar 17 09:28:00 crc kubenswrapper[4813]: I0317 09:28:00.456549 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562328-pmv66" Mar 17 09:28:00 crc kubenswrapper[4813]: I0317 09:28:00.548814 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-metrics-certs\") pod \"openstack-operator-controller-manager-65f8859ffd-mxfzh\" (UID: \"db193bcc-748f-4830-873c-48a2f9229b40\") " pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:28:00 crc kubenswrapper[4813]: I0317 09:28:00.548983 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-webhook-certs\") pod \"openstack-operator-controller-manager-65f8859ffd-mxfzh\" (UID: \"db193bcc-748f-4830-873c-48a2f9229b40\") " pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:28:00 crc kubenswrapper[4813]: E0317 09:28:00.549011 4813 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Mar 17 09:28:00 crc kubenswrapper[4813]: E0317 09:28:00.549087 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-metrics-certs podName:db193bcc-748f-4830-873c-48a2f9229b40 nodeName:}" failed. No retries permitted until 2026-03-17 09:28:16.549070079 +0000 UTC m=+1118.649873578 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-metrics-certs") pod "openstack-operator-controller-manager-65f8859ffd-mxfzh" (UID: "db193bcc-748f-4830-873c-48a2f9229b40") : secret "metrics-server-cert" not found Mar 17 09:28:00 crc kubenswrapper[4813]: E0317 09:28:00.549090 4813 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Mar 17 09:28:00 crc kubenswrapper[4813]: E0317 09:28:00.549128 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-webhook-certs podName:db193bcc-748f-4830-873c-48a2f9229b40 nodeName:}" failed. No retries permitted until 2026-03-17 09:28:16.549118511 +0000 UTC m=+1118.649922010 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-webhook-certs") pod "openstack-operator-controller-manager-65f8859ffd-mxfzh" (UID: "db193bcc-748f-4830-873c-48a2f9229b40") : secret "webhook-server-cert" not found Mar 17 09:28:04 crc kubenswrapper[4813]: I0317 09:28:04.018474 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-59bc569d95-xkfkm" Mar 17 09:28:04 crc kubenswrapper[4813]: I0317 09:28:04.020883 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-8d58dc466-9cqth" Mar 17 09:28:04 crc kubenswrapper[4813]: I0317 09:28:04.041712 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-588d4d986b-tn4pv" Mar 17 09:28:04 crc kubenswrapper[4813]: I0317 09:28:04.052280 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-79df6bcc97-wzmxx" Mar 17 09:28:04 crc kubenswrapper[4813]: I0317 09:28:04.192878 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-67dd5f86f5-hng2s" Mar 17 09:28:04 crc kubenswrapper[4813]: I0317 09:28:04.295622 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-7965b" event={"ID":"c9ea4c59-d87b-4e58-ac8f-2786f8f32f02","Type":"ContainerStarted","Data":"34654678b3d6bb9e6a409e642cbb5c7f7f1875bf0540e7f6c96989cdbadf29a4"} Mar 17 09:28:04 crc kubenswrapper[4813]: I0317 09:28:04.298324 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-67ccfc9778-ngd7n" Mar 17 09:28:04 crc kubenswrapper[4813]: I0317 09:28:04.380138 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-768b96df4c-c966n" Mar 17 09:28:04 crc kubenswrapper[4813]: I0317 09:28:04.418937 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562328-pmv66"] Mar 17 09:28:04 crc kubenswrapper[4813]: I0317 09:28:04.465691 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-55f864c847-dbs68" Mar 17 09:28:04 crc kubenswrapper[4813]: I0317 09:28:04.511732 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-767865f676-mnsb4" Mar 17 09:28:04 crc kubenswrapper[4813]: I0317 09:28:04.521590 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6f787dddc9-4ds8c" Mar 17 09:28:04 crc kubenswrapper[4813]: I0317 09:28:04.643697 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-5b9f45d989-mps6t" Mar 17 09:28:04 crc kubenswrapper[4813]: I0317 09:28:04.721990 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-884679f54-7d66d" Mar 17 09:28:05 crc kubenswrapper[4813]: W0317 09:28:05.031492 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod02f74478_75b1_4e8a_b347_d0bc8cc32f4c.slice/crio-7ab2ac236a8a945df78f6cde8e02e278a3ca69ade5ce09b4b85828cef90e3a57 WatchSource:0}: Error finding container 7ab2ac236a8a945df78f6cde8e02e278a3ca69ade5ce09b4b85828cef90e3a57: Status 404 returned error can't find the container with id 7ab2ac236a8a945df78f6cde8e02e278a3ca69ade5ce09b4b85828cef90e3a57 Mar 17 09:28:05 crc kubenswrapper[4813]: I0317 09:28:05.302997 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562328-pmv66" event={"ID":"02f74478-75b1-4e8a-b347-d0bc8cc32f4c","Type":"ContainerStarted","Data":"7ab2ac236a8a945df78f6cde8e02e278a3ca69ade5ce09b4b85828cef90e3a57"} Mar 17 09:28:09 crc kubenswrapper[4813]: I0317 09:28:09.315782 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-sql9s" Mar 17 09:28:09 crc kubenswrapper[4813]: I0317 09:28:09.316150 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-sql9s" Mar 17 09:28:09 crc kubenswrapper[4813]: I0317 09:28:09.397085 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-sql9s" Mar 17 09:28:09 crc kubenswrapper[4813]: I0317 09:28:09.468676 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-sql9s" Mar 17 09:28:09 crc kubenswrapper[4813]: I0317 09:28:09.645949 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sql9s"] Mar 17 09:28:11 crc kubenswrapper[4813]: I0317 09:28:11.353390 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-sql9s" podUID="19be6482-9c7f-465b-8a3d-5b99669fa87d" containerName="registry-server" containerID="cri-o://2b0c94d67f86b72c7cd559d403934130305882ed950f70faf934bbcd8fe89097" gracePeriod=2 Mar 17 09:28:12 crc kubenswrapper[4813]: I0317 09:28:12.368804 4813 generic.go:334] "Generic (PLEG): container finished" podID="19be6482-9c7f-465b-8a3d-5b99669fa87d" containerID="2b0c94d67f86b72c7cd559d403934130305882ed950f70faf934bbcd8fe89097" exitCode=0 Mar 17 09:28:12 crc kubenswrapper[4813]: I0317 09:28:12.368842 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sql9s" event={"ID":"19be6482-9c7f-465b-8a3d-5b99669fa87d","Type":"ContainerDied","Data":"2b0c94d67f86b72c7cd559d403934130305882ed950f70faf934bbcd8fe89097"} Mar 17 09:28:15 crc kubenswrapper[4813]: E0317 09:28:15.677661 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42" Mar 17 09:28:15 crc kubenswrapper[4813]: E0317 09:28:15.678546 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zv8d6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5c5cb9c4d7-r2ksk_openstack-operators(860ccadc-ca76-4523-9543-2b0960eb77a9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 09:28:15 crc kubenswrapper[4813]: E0317 09:28:15.679717 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-r2ksk" podUID="860ccadc-ca76-4523-9543-2b0960eb77a9" Mar 17 09:28:15 crc kubenswrapper[4813]: I0317 09:28:15.715559 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sql9s" Mar 17 09:28:15 crc kubenswrapper[4813]: I0317 09:28:15.893148 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xnqb\" (UniqueName: \"kubernetes.io/projected/19be6482-9c7f-465b-8a3d-5b99669fa87d-kube-api-access-8xnqb\") pod \"19be6482-9c7f-465b-8a3d-5b99669fa87d\" (UID: \"19be6482-9c7f-465b-8a3d-5b99669fa87d\") " Mar 17 09:28:15 crc kubenswrapper[4813]: I0317 09:28:15.893194 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19be6482-9c7f-465b-8a3d-5b99669fa87d-catalog-content\") pod \"19be6482-9c7f-465b-8a3d-5b99669fa87d\" (UID: \"19be6482-9c7f-465b-8a3d-5b99669fa87d\") " Mar 17 09:28:15 crc kubenswrapper[4813]: I0317 09:28:15.893281 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19be6482-9c7f-465b-8a3d-5b99669fa87d-utilities\") pod \"19be6482-9c7f-465b-8a3d-5b99669fa87d\" (UID: \"19be6482-9c7f-465b-8a3d-5b99669fa87d\") " Mar 17 09:28:15 crc kubenswrapper[4813]: I0317 09:28:15.894484 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19be6482-9c7f-465b-8a3d-5b99669fa87d-utilities" (OuterVolumeSpecName: "utilities") pod "19be6482-9c7f-465b-8a3d-5b99669fa87d" (UID: "19be6482-9c7f-465b-8a3d-5b99669fa87d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:28:15 crc kubenswrapper[4813]: I0317 09:28:15.900187 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19be6482-9c7f-465b-8a3d-5b99669fa87d-kube-api-access-8xnqb" (OuterVolumeSpecName: "kube-api-access-8xnqb") pod "19be6482-9c7f-465b-8a3d-5b99669fa87d" (UID: "19be6482-9c7f-465b-8a3d-5b99669fa87d"). InnerVolumeSpecName "kube-api-access-8xnqb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:28:15 crc kubenswrapper[4813]: I0317 09:28:15.916402 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19be6482-9c7f-465b-8a3d-5b99669fa87d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "19be6482-9c7f-465b-8a3d-5b99669fa87d" (UID: "19be6482-9c7f-465b-8a3d-5b99669fa87d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:28:15 crc kubenswrapper[4813]: I0317 09:28:15.995265 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xnqb\" (UniqueName: \"kubernetes.io/projected/19be6482-9c7f-465b-8a3d-5b99669fa87d-kube-api-access-8xnqb\") on node \"crc\" DevicePath \"\"" Mar 17 09:28:15 crc kubenswrapper[4813]: I0317 09:28:15.995330 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19be6482-9c7f-465b-8a3d-5b99669fa87d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:28:15 crc kubenswrapper[4813]: I0317 09:28:15.995357 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19be6482-9c7f-465b-8a3d-5b99669fa87d-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:28:16 crc kubenswrapper[4813]: E0317 09:28:16.150918 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622" Mar 17 09:28:16 crc kubenswrapper[4813]: E0317 09:28:16.151080 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jnfhn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-5784578c99-hcsvd_openstack-operators(968158c3-2dfc-439e-9d20-8706d611fec1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 09:28:16 crc kubenswrapper[4813]: E0317 09:28:16.152296 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-hcsvd" podUID="968158c3-2dfc-439e-9d20-8706d611fec1" Mar 17 09:28:16 crc kubenswrapper[4813]: I0317 09:28:16.323025 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8e80dcf7-dae9-4faa-ab3f-b547b570926b-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-g5vgd\" (UID: \"8e80dcf7-dae9-4faa-ab3f-b547b570926b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" Mar 17 09:28:16 crc kubenswrapper[4813]: I0317 09:28:16.332342 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/8e80dcf7-dae9-4faa-ab3f-b547b570926b-cert\") pod \"openstack-baremetal-operator-controller-manager-89d64c458-g5vgd\" (UID: \"8e80dcf7-dae9-4faa-ab3f-b547b570926b\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" Mar 17 09:28:16 crc kubenswrapper[4813]: I0317 09:28:16.413811 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-sql9s" event={"ID":"19be6482-9c7f-465b-8a3d-5b99669fa87d","Type":"ContainerDied","Data":"15ea27d29d16b3f457ec3fb7c97da716ced37ad5022a0378ed44d37c2167af51"} Mar 17 09:28:16 crc kubenswrapper[4813]: I0317 09:28:16.413895 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-sql9s" Mar 17 09:28:16 crc kubenswrapper[4813]: I0317 09:28:16.413923 4813 scope.go:117] "RemoveContainer" containerID="2b0c94d67f86b72c7cd559d403934130305882ed950f70faf934bbcd8fe89097" Mar 17 09:28:16 crc kubenswrapper[4813]: I0317 09:28:16.442498 4813 scope.go:117] "RemoveContainer" containerID="8762c2aa69313afeacf3cda72bbc3bfff53d87f94d3acbca115e5c7eae6abaf9" Mar 17 09:28:16 crc kubenswrapper[4813]: I0317 09:28:16.454085 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-sql9s"] Mar 17 09:28:16 crc kubenswrapper[4813]: I0317 09:28:16.462119 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-sql9s"] Mar 17 09:28:16 crc kubenswrapper[4813]: I0317 09:28:16.490106 4813 scope.go:117] "RemoveContainer" containerID="68eeb55595f33e79510e97fd8c2eb0064317b08a9de1a31c9069261af29e521a" Mar 17 09:28:16 crc kubenswrapper[4813]: I0317 09:28:16.551716 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" Mar 17 09:28:16 crc kubenswrapper[4813]: I0317 09:28:16.630075 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-webhook-certs\") pod \"openstack-operator-controller-manager-65f8859ffd-mxfzh\" (UID: \"db193bcc-748f-4830-873c-48a2f9229b40\") " pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:28:16 crc kubenswrapper[4813]: I0317 09:28:16.630136 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-metrics-certs\") pod \"openstack-operator-controller-manager-65f8859ffd-mxfzh\" (UID: \"db193bcc-748f-4830-873c-48a2f9229b40\") " pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:28:16 crc kubenswrapper[4813]: I0317 09:28:16.634072 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-metrics-certs\") pod \"openstack-operator-controller-manager-65f8859ffd-mxfzh\" (UID: \"db193bcc-748f-4830-873c-48a2f9229b40\") " pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:28:16 crc kubenswrapper[4813]: I0317 09:28:16.634650 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/db193bcc-748f-4830-873c-48a2f9229b40-webhook-certs\") pod \"openstack-operator-controller-manager-65f8859ffd-mxfzh\" (UID: \"db193bcc-748f-4830-873c-48a2f9229b40\") " pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:28:16 crc kubenswrapper[4813]: I0317 09:28:16.733366 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:28:16 crc kubenswrapper[4813]: I0317 09:28:16.751970 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19be6482-9c7f-465b-8a3d-5b99669fa87d" path="/var/lib/kubelet/pods/19be6482-9c7f-465b-8a3d-5b99669fa87d/volumes" Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.030882 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd"] Mar 17 09:28:17 crc kubenswrapper[4813]: W0317 09:28:17.032309 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8e80dcf7_dae9_4faa_ab3f_b547b570926b.slice/crio-c3a4cc5a0bbd9384189486f297e276617e3780de353d3b3babc1bebb165272c2 WatchSource:0}: Error finding container c3a4cc5a0bbd9384189486f297e276617e3780de353d3b3babc1bebb165272c2: Status 404 returned error can't find the container with id c3a4cc5a0bbd9384189486f297e276617e3780de353d3b3babc1bebb165272c2 Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.242447 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh"] Mar 17 09:28:17 crc kubenswrapper[4813]: W0317 09:28:17.249319 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddb193bcc_748f_4830_873c_48a2f9229b40.slice/crio-4ea08633af5b128d6b2ec2ba5b9eaee3d980eaab19cd1260a39f4c4cb012b8b6 WatchSource:0}: Error finding container 4ea08633af5b128d6b2ec2ba5b9eaee3d980eaab19cd1260a39f4c4cb012b8b6: Status 404 returned error can't find the container with id 4ea08633af5b128d6b2ec2ba5b9eaee3d980eaab19cd1260a39f4c4cb012b8b6 Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.432041 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-52xx2" event={"ID":"ec48dbf8-730b-4a2e-9986-03d8f66de013","Type":"ContainerStarted","Data":"4e4ddbf2053a4d51fc5cd8b8213e0f4fd1c2c97985ef93331681b487e06f4395"} Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.432205 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-52xx2" Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.433580 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-7965b" event={"ID":"c9ea4c59-d87b-4e58-ac8f-2786f8f32f02","Type":"ContainerStarted","Data":"594f464dd412a35434f36ef3873e58e4637aa1a3b6f2b528f0d9460d6c6e9f97"} Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.433949 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-7965b" Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.435698 4813 generic.go:334] "Generic (PLEG): container finished" podID="02f74478-75b1-4e8a-b347-d0bc8cc32f4c" containerID="f60ccc154e2565faf3d3ff26f1d684ed650435d011b99faac17cdf34034d77c9" exitCode=0 Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.435809 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562328-pmv66" event={"ID":"02f74478-75b1-4e8a-b347-d0bc8cc32f4c","Type":"ContainerDied","Data":"f60ccc154e2565faf3d3ff26f1d684ed650435d011b99faac17cdf34034d77c9"} Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.437088 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tqg57" event={"ID":"b4b9a61e-819f-4da5-a28a-8e49c0a3f7cd","Type":"ContainerStarted","Data":"89e7328db19559ada6e106c4c138eba270c24bc4231348732c14855577a26e20"} Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.443801 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tqg57" Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.453351 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-52xx2" podStartSLOduration=3.4925419 podStartE2EDuration="33.453333688s" podCreationTimestamp="2026-03-17 09:27:44 +0000 UTC" firstStartedPulling="2026-03-17 09:27:45.713837234 +0000 UTC m=+1087.814640733" lastFinishedPulling="2026-03-17 09:28:15.674629022 +0000 UTC m=+1117.775432521" observedRunningTime="2026-03-17 09:28:17.453152412 +0000 UTC m=+1119.553955941" watchObservedRunningTime="2026-03-17 09:28:17.453333688 +0000 UTC m=+1119.554137187" Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.455247 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-tfhsc" event={"ID":"668691cc-39f3-4b04-8ed6-bc86f9965f12","Type":"ContainerStarted","Data":"a031ab1b1aa4fa9a79686e4f68c2071ad60eaa597ed38d1b2a080cb5af3dd86a"} Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.456219 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-tfhsc" Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.460963 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rxzjs" event={"ID":"bcd2954f-b240-4cc3-8387-15b0e64a9721","Type":"ContainerStarted","Data":"243abb56764436ced05a4cc3046599cf161e08083ac1efb4ac8a5bfa03c326f0"} Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.463683 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" event={"ID":"db193bcc-748f-4830-873c-48a2f9229b40","Type":"ContainerStarted","Data":"282be40ea5774e241f3c6acde35c74593197690bae03ea787495afefbc4710a3"} Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.463726 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" event={"ID":"db193bcc-748f-4830-873c-48a2f9229b40","Type":"ContainerStarted","Data":"4ea08633af5b128d6b2ec2ba5b9eaee3d980eaab19cd1260a39f4c4cb012b8b6"} Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.463829 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.465146 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-wrtl9" event={"ID":"520d3ab3-78d7-4be3-8466-9dadfce15c2c","Type":"ContainerStarted","Data":"5859ef0136f9c762d7a39f22a346be68c8e0bf8ffd4e90fa21d127a706505f7f"} Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.465312 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-wrtl9" Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.468168 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-c674c5965-njxx7" event={"ID":"ac052585-f733-47d5-91b9-87ad7957e511","Type":"ContainerStarted","Data":"2c27eb6cfa0c1049f8c18fa4f3e1b194cb5ff6593236d291e7c44d39ac345382"} Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.468407 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-c674c5965-njxx7" Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.469248 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" event={"ID":"8e80dcf7-dae9-4faa-ab3f-b547b570926b","Type":"ContainerStarted","Data":"c3a4cc5a0bbd9384189486f297e276617e3780de353d3b3babc1bebb165272c2"} Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.505406 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-7965b" podStartSLOduration=21.680656304 podStartE2EDuration="34.505385993s" podCreationTimestamp="2026-03-17 09:27:43 +0000 UTC" firstStartedPulling="2026-03-17 09:28:03.332817491 +0000 UTC m=+1105.433620990" lastFinishedPulling="2026-03-17 09:28:16.15754718 +0000 UTC m=+1118.258350679" observedRunningTime="2026-03-17 09:28:17.499923581 +0000 UTC m=+1119.600727100" watchObservedRunningTime="2026-03-17 09:28:17.505385993 +0000 UTC m=+1119.606189492" Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.517786 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tqg57" podStartSLOduration=3.5676302680000003 podStartE2EDuration="33.517766002s" podCreationTimestamp="2026-03-17 09:27:44 +0000 UTC" firstStartedPulling="2026-03-17 09:27:45.713756261 +0000 UTC m=+1087.814559760" lastFinishedPulling="2026-03-17 09:28:15.663891985 +0000 UTC m=+1117.764695494" observedRunningTime="2026-03-17 09:28:17.516902134 +0000 UTC m=+1119.617705643" watchObservedRunningTime="2026-03-17 09:28:17.517766002 +0000 UTC m=+1119.618569501" Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.538727 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-c674c5965-njxx7" podStartSLOduration=4.236245008 podStartE2EDuration="33.53870918s" podCreationTimestamp="2026-03-17 09:27:44 +0000 UTC" firstStartedPulling="2026-03-17 09:27:45.642277567 +0000 UTC m=+1087.743081066" lastFinishedPulling="2026-03-17 09:28:14.944741699 +0000 UTC m=+1117.045545238" observedRunningTime="2026-03-17 09:28:17.53584486 +0000 UTC m=+1119.636648369" watchObservedRunningTime="2026-03-17 09:28:17.53870918 +0000 UTC m=+1119.639512679" Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.562462 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-wrtl9" podStartSLOduration=4.331742387 podStartE2EDuration="33.562441665s" podCreationTimestamp="2026-03-17 09:27:44 +0000 UTC" firstStartedPulling="2026-03-17 09:27:45.714001009 +0000 UTC m=+1087.814804508" lastFinishedPulling="2026-03-17 09:28:14.944700257 +0000 UTC m=+1117.045503786" observedRunningTime="2026-03-17 09:28:17.553264367 +0000 UTC m=+1119.654067866" watchObservedRunningTime="2026-03-17 09:28:17.562441665 +0000 UTC m=+1119.663245184" Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.593346 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" podStartSLOduration=33.593327215 podStartE2EDuration="33.593327215s" podCreationTimestamp="2026-03-17 09:27:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:28:17.590513097 +0000 UTC m=+1119.691316596" watchObservedRunningTime="2026-03-17 09:28:17.593327215 +0000 UTC m=+1119.694130754" Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.614379 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-rxzjs" podStartSLOduration=3.653976891 podStartE2EDuration="33.614362576s" podCreationTimestamp="2026-03-17 09:27:44 +0000 UTC" firstStartedPulling="2026-03-17 09:27:45.714434573 +0000 UTC m=+1087.815238072" lastFinishedPulling="2026-03-17 09:28:15.674820258 +0000 UTC m=+1117.775623757" observedRunningTime="2026-03-17 09:28:17.609388519 +0000 UTC m=+1119.710192018" watchObservedRunningTime="2026-03-17 09:28:17.614362576 +0000 UTC m=+1119.715166075" Mar 17 09:28:17 crc kubenswrapper[4813]: I0317 09:28:17.629944 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-tfhsc" podStartSLOduration=5.238038103 podStartE2EDuration="34.629925805s" podCreationTimestamp="2026-03-17 09:27:43 +0000 UTC" firstStartedPulling="2026-03-17 09:27:45.55420935 +0000 UTC m=+1087.655012849" lastFinishedPulling="2026-03-17 09:28:14.946097022 +0000 UTC m=+1117.046900551" observedRunningTime="2026-03-17 09:28:17.625644751 +0000 UTC m=+1119.726448250" watchObservedRunningTime="2026-03-17 09:28:17.629925805 +0000 UTC m=+1119.730729324" Mar 17 09:28:19 crc kubenswrapper[4813]: I0317 09:28:19.080618 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562328-pmv66" Mar 17 09:28:19 crc kubenswrapper[4813]: I0317 09:28:19.205700 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g27vx\" (UniqueName: \"kubernetes.io/projected/02f74478-75b1-4e8a-b347-d0bc8cc32f4c-kube-api-access-g27vx\") pod \"02f74478-75b1-4e8a-b347-d0bc8cc32f4c\" (UID: \"02f74478-75b1-4e8a-b347-d0bc8cc32f4c\") " Mar 17 09:28:19 crc kubenswrapper[4813]: I0317 09:28:19.216166 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02f74478-75b1-4e8a-b347-d0bc8cc32f4c-kube-api-access-g27vx" (OuterVolumeSpecName: "kube-api-access-g27vx") pod "02f74478-75b1-4e8a-b347-d0bc8cc32f4c" (UID: "02f74478-75b1-4e8a-b347-d0bc8cc32f4c"). InnerVolumeSpecName "kube-api-access-g27vx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:28:19 crc kubenswrapper[4813]: I0317 09:28:19.307355 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g27vx\" (UniqueName: \"kubernetes.io/projected/02f74478-75b1-4e8a-b347-d0bc8cc32f4c-kube-api-access-g27vx\") on node \"crc\" DevicePath \"\"" Mar 17 09:28:19 crc kubenswrapper[4813]: I0317 09:28:19.506942 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562328-pmv66" event={"ID":"02f74478-75b1-4e8a-b347-d0bc8cc32f4c","Type":"ContainerDied","Data":"7ab2ac236a8a945df78f6cde8e02e278a3ca69ade5ce09b4b85828cef90e3a57"} Mar 17 09:28:19 crc kubenswrapper[4813]: I0317 09:28:19.507021 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ab2ac236a8a945df78f6cde8e02e278a3ca69ade5ce09b4b85828cef90e3a57" Mar 17 09:28:19 crc kubenswrapper[4813]: I0317 09:28:19.506941 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562328-pmv66" Mar 17 09:28:19 crc kubenswrapper[4813]: I0317 09:28:19.511449 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" event={"ID":"8e80dcf7-dae9-4faa-ab3f-b547b570926b","Type":"ContainerStarted","Data":"e5d3f46f9761c27b20772a103c35c33210bd1dbdc43a2c16c96a514307b79a3b"} Mar 17 09:28:19 crc kubenswrapper[4813]: I0317 09:28:19.512336 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" Mar 17 09:28:19 crc kubenswrapper[4813]: I0317 09:28:19.559784 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" podStartSLOduration=33.52120729 podStartE2EDuration="35.559763747s" podCreationTimestamp="2026-03-17 09:27:44 +0000 UTC" firstStartedPulling="2026-03-17 09:28:17.034358879 +0000 UTC m=+1119.135162378" lastFinishedPulling="2026-03-17 09:28:19.072915336 +0000 UTC m=+1121.173718835" observedRunningTime="2026-03-17 09:28:19.557831886 +0000 UTC m=+1121.658635395" watchObservedRunningTime="2026-03-17 09:28:19.559763747 +0000 UTC m=+1121.660567256" Mar 17 09:28:20 crc kubenswrapper[4813]: I0317 09:28:20.146432 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562322-q8wzc"] Mar 17 09:28:20 crc kubenswrapper[4813]: I0317 09:28:20.152680 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562322-q8wzc"] Mar 17 09:28:20 crc kubenswrapper[4813]: I0317 09:28:20.741044 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a3fed28-e266-47a5-851f-f04fb94804e8" path="/var/lib/kubelet/pods/6a3fed28-e266-47a5-851f-f04fb94804e8/volumes" Mar 17 09:28:24 crc kubenswrapper[4813]: I0317 09:28:24.459571 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-8464cc45fb-tfhsc" Mar 17 09:28:24 crc kubenswrapper[4813]: I0317 09:28:24.617375 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-5d488d59fb-tqg57" Mar 17 09:28:24 crc kubenswrapper[4813]: I0317 09:28:24.788506 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-c674c5965-njxx7" Mar 17 09:28:24 crc kubenswrapper[4813]: I0317 09:28:24.811837 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-d6b694c5-52xx2" Mar 17 09:28:24 crc kubenswrapper[4813]: I0317 09:28:24.913202 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-6c4d75f7f9-wrtl9" Mar 17 09:28:26 crc kubenswrapper[4813]: I0317 09:28:26.563739 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-89d64c458-g5vgd" Mar 17 09:28:26 crc kubenswrapper[4813]: I0317 09:28:26.740888 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-65f8859ffd-mxfzh" Mar 17 09:28:29 crc kubenswrapper[4813]: E0317 09:28:29.733375 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:43bd420bc05b4789243740bc75f61e10c7aac7883fc2f82b2d4d50085bc96c42\\\"\"" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-r2ksk" podUID="860ccadc-ca76-4523-9543-2b0960eb77a9" Mar 17 09:28:29 crc kubenswrapper[4813]: I0317 09:28:29.867830 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-7b9c774f96-7965b" Mar 17 09:28:30 crc kubenswrapper[4813]: E0317 09:28:30.731750 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:c8743a6661d118b0e5ba3eb110643358a8a3237dc75984a8f9829880b55a1622\\\"\"" pod="openstack-operators/placement-operator-controller-manager-5784578c99-hcsvd" podUID="968158c3-2dfc-439e-9d20-8706d611fec1" Mar 17 09:28:39 crc kubenswrapper[4813]: I0317 09:28:39.740980 4813 scope.go:117] "RemoveContainer" containerID="7213f79e91ecdab9ed18fc3a07525828594cef72a9583cc67db80f37da5f05eb" Mar 17 09:28:43 crc kubenswrapper[4813]: I0317 09:28:43.717404 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-r2ksk" event={"ID":"860ccadc-ca76-4523-9543-2b0960eb77a9","Type":"ContainerStarted","Data":"696e82f9d8ef8a9df65901b06f3f0ab87d18aab1d2b2cd921f10457d73dc0fbd"} Mar 17 09:28:43 crc kubenswrapper[4813]: I0317 09:28:43.718133 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-r2ksk" Mar 17 09:28:43 crc kubenswrapper[4813]: I0317 09:28:43.739451 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-r2ksk" podStartSLOduration=2.454200037 podStartE2EDuration="59.739429974s" podCreationTimestamp="2026-03-17 09:27:44 +0000 UTC" firstStartedPulling="2026-03-17 09:27:45.641849913 +0000 UTC m=+1087.742653412" lastFinishedPulling="2026-03-17 09:28:42.92707981 +0000 UTC m=+1145.027883349" observedRunningTime="2026-03-17 09:28:43.73676319 +0000 UTC m=+1145.837566719" watchObservedRunningTime="2026-03-17 09:28:43.739429974 +0000 UTC m=+1145.840233503" Mar 17 09:28:46 crc kubenswrapper[4813]: I0317 09:28:46.763029 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-5784578c99-hcsvd" event={"ID":"968158c3-2dfc-439e-9d20-8706d611fec1","Type":"ContainerStarted","Data":"ce4de4b6b1f7f7f9fc609db6990f1f9793057efdb4b8aa7e85b8064e2e8ac8e7"} Mar 17 09:28:46 crc kubenswrapper[4813]: I0317 09:28:46.765988 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-5784578c99-hcsvd" Mar 17 09:28:46 crc kubenswrapper[4813]: I0317 09:28:46.796526 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-5784578c99-hcsvd" podStartSLOduration=2.179970445 podStartE2EDuration="1m2.796497301s" podCreationTimestamp="2026-03-17 09:27:44 +0000 UTC" firstStartedPulling="2026-03-17 09:27:45.653471698 +0000 UTC m=+1087.754275207" lastFinishedPulling="2026-03-17 09:28:46.269998544 +0000 UTC m=+1148.370802063" observedRunningTime="2026-03-17 09:28:46.788484369 +0000 UTC m=+1148.889287898" watchObservedRunningTime="2026-03-17 09:28:46.796497301 +0000 UTC m=+1148.897300840" Mar 17 09:28:54 crc kubenswrapper[4813]: I0317 09:28:54.665774 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-5784578c99-hcsvd" Mar 17 09:28:54 crc kubenswrapper[4813]: I0317 09:28:54.837002 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5c5cb9c4d7-r2ksk" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.721730 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-988sz"] Mar 17 09:29:11 crc kubenswrapper[4813]: E0317 09:29:11.722490 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19be6482-9c7f-465b-8a3d-5b99669fa87d" containerName="extract-utilities" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.722504 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="19be6482-9c7f-465b-8a3d-5b99669fa87d" containerName="extract-utilities" Mar 17 09:29:11 crc kubenswrapper[4813]: E0317 09:29:11.722523 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19be6482-9c7f-465b-8a3d-5b99669fa87d" containerName="registry-server" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.722531 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="19be6482-9c7f-465b-8a3d-5b99669fa87d" containerName="registry-server" Mar 17 09:29:11 crc kubenswrapper[4813]: E0317 09:29:11.722555 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="19be6482-9c7f-465b-8a3d-5b99669fa87d" containerName="extract-content" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.722563 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="19be6482-9c7f-465b-8a3d-5b99669fa87d" containerName="extract-content" Mar 17 09:29:11 crc kubenswrapper[4813]: E0317 09:29:11.722571 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02f74478-75b1-4e8a-b347-d0bc8cc32f4c" containerName="oc" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.722576 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="02f74478-75b1-4e8a-b347-d0bc8cc32f4c" containerName="oc" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.722707 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="19be6482-9c7f-465b-8a3d-5b99669fa87d" containerName="registry-server" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.722721 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="02f74478-75b1-4e8a-b347-d0bc8cc32f4c" containerName="oc" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.723400 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-988sz" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.726540 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.726817 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.727082 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-6lqtx" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.728021 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.737844 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-988sz"] Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.830303 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z9lkn"] Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.837966 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-z9lkn" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.858555 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.864839 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z9lkn"] Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.883100 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63458ae3-263a-4d68-95bd-4824acb4cebd-config\") pod \"dnsmasq-dns-675f4bcbfc-988sz\" (UID: \"63458ae3-263a-4d68-95bd-4824acb4cebd\") " pod="openstack/dnsmasq-dns-675f4bcbfc-988sz" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.883346 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzgpg\" (UniqueName: \"kubernetes.io/projected/63458ae3-263a-4d68-95bd-4824acb4cebd-kube-api-access-lzgpg\") pod \"dnsmasq-dns-675f4bcbfc-988sz\" (UID: \"63458ae3-263a-4d68-95bd-4824acb4cebd\") " pod="openstack/dnsmasq-dns-675f4bcbfc-988sz" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.984218 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqrfw\" (UniqueName: \"kubernetes.io/projected/8369f72b-14fd-40f4-864c-c8bddd0f43e8-kube-api-access-sqrfw\") pod \"dnsmasq-dns-78dd6ddcc-z9lkn\" (UID: \"8369f72b-14fd-40f4-864c-c8bddd0f43e8\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z9lkn" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.984277 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8369f72b-14fd-40f4-864c-c8bddd0f43e8-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-z9lkn\" (UID: \"8369f72b-14fd-40f4-864c-c8bddd0f43e8\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z9lkn" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.984308 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzgpg\" (UniqueName: \"kubernetes.io/projected/63458ae3-263a-4d68-95bd-4824acb4cebd-kube-api-access-lzgpg\") pod \"dnsmasq-dns-675f4bcbfc-988sz\" (UID: \"63458ae3-263a-4d68-95bd-4824acb4cebd\") " pod="openstack/dnsmasq-dns-675f4bcbfc-988sz" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.984440 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8369f72b-14fd-40f4-864c-c8bddd0f43e8-config\") pod \"dnsmasq-dns-78dd6ddcc-z9lkn\" (UID: \"8369f72b-14fd-40f4-864c-c8bddd0f43e8\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z9lkn" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.984570 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63458ae3-263a-4d68-95bd-4824acb4cebd-config\") pod \"dnsmasq-dns-675f4bcbfc-988sz\" (UID: \"63458ae3-263a-4d68-95bd-4824acb4cebd\") " pod="openstack/dnsmasq-dns-675f4bcbfc-988sz" Mar 17 09:29:11 crc kubenswrapper[4813]: I0317 09:29:11.985469 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63458ae3-263a-4d68-95bd-4824acb4cebd-config\") pod \"dnsmasq-dns-675f4bcbfc-988sz\" (UID: \"63458ae3-263a-4d68-95bd-4824acb4cebd\") " pod="openstack/dnsmasq-dns-675f4bcbfc-988sz" Mar 17 09:29:12 crc kubenswrapper[4813]: I0317 09:29:12.008048 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzgpg\" (UniqueName: \"kubernetes.io/projected/63458ae3-263a-4d68-95bd-4824acb4cebd-kube-api-access-lzgpg\") pod \"dnsmasq-dns-675f4bcbfc-988sz\" (UID: \"63458ae3-263a-4d68-95bd-4824acb4cebd\") " pod="openstack/dnsmasq-dns-675f4bcbfc-988sz" Mar 17 09:29:12 crc kubenswrapper[4813]: I0317 09:29:12.072989 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-988sz" Mar 17 09:29:12 crc kubenswrapper[4813]: I0317 09:29:12.085587 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8369f72b-14fd-40f4-864c-c8bddd0f43e8-config\") pod \"dnsmasq-dns-78dd6ddcc-z9lkn\" (UID: \"8369f72b-14fd-40f4-864c-c8bddd0f43e8\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z9lkn" Mar 17 09:29:12 crc kubenswrapper[4813]: I0317 09:29:12.085704 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqrfw\" (UniqueName: \"kubernetes.io/projected/8369f72b-14fd-40f4-864c-c8bddd0f43e8-kube-api-access-sqrfw\") pod \"dnsmasq-dns-78dd6ddcc-z9lkn\" (UID: \"8369f72b-14fd-40f4-864c-c8bddd0f43e8\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z9lkn" Mar 17 09:29:12 crc kubenswrapper[4813]: I0317 09:29:12.085742 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8369f72b-14fd-40f4-864c-c8bddd0f43e8-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-z9lkn\" (UID: \"8369f72b-14fd-40f4-864c-c8bddd0f43e8\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z9lkn" Mar 17 09:29:12 crc kubenswrapper[4813]: I0317 09:29:12.086444 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8369f72b-14fd-40f4-864c-c8bddd0f43e8-config\") pod \"dnsmasq-dns-78dd6ddcc-z9lkn\" (UID: \"8369f72b-14fd-40f4-864c-c8bddd0f43e8\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z9lkn" Mar 17 09:29:12 crc kubenswrapper[4813]: I0317 09:29:12.086689 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8369f72b-14fd-40f4-864c-c8bddd0f43e8-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-z9lkn\" (UID: \"8369f72b-14fd-40f4-864c-c8bddd0f43e8\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z9lkn" Mar 17 09:29:12 crc kubenswrapper[4813]: I0317 09:29:12.114486 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqrfw\" (UniqueName: \"kubernetes.io/projected/8369f72b-14fd-40f4-864c-c8bddd0f43e8-kube-api-access-sqrfw\") pod \"dnsmasq-dns-78dd6ddcc-z9lkn\" (UID: \"8369f72b-14fd-40f4-864c-c8bddd0f43e8\") " pod="openstack/dnsmasq-dns-78dd6ddcc-z9lkn" Mar 17 09:29:12 crc kubenswrapper[4813]: I0317 09:29:12.174897 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-z9lkn" Mar 17 09:29:12 crc kubenswrapper[4813]: I0317 09:29:12.475761 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-988sz"] Mar 17 09:29:12 crc kubenswrapper[4813]: I0317 09:29:12.588096 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z9lkn"] Mar 17 09:29:12 crc kubenswrapper[4813]: W0317 09:29:12.591171 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8369f72b_14fd_40f4_864c_c8bddd0f43e8.slice/crio-fb168c599dae50c50502a8d04510e9177e261e111ec9100752b12a727393c6bb WatchSource:0}: Error finding container fb168c599dae50c50502a8d04510e9177e261e111ec9100752b12a727393c6bb: Status 404 returned error can't find the container with id fb168c599dae50c50502a8d04510e9177e261e111ec9100752b12a727393c6bb Mar 17 09:29:12 crc kubenswrapper[4813]: I0317 09:29:12.992384 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-988sz" event={"ID":"63458ae3-263a-4d68-95bd-4824acb4cebd","Type":"ContainerStarted","Data":"a33ce5fe3932d6af3b62fc277032cd57b216275da6889e83b581d90b67bcfa9c"} Mar 17 09:29:12 crc kubenswrapper[4813]: I0317 09:29:12.995113 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-z9lkn" event={"ID":"8369f72b-14fd-40f4-864c-c8bddd0f43e8","Type":"ContainerStarted","Data":"fb168c599dae50c50502a8d04510e9177e261e111ec9100752b12a727393c6bb"} Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.113609 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.113925 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.553215 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-988sz"] Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.572281 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-6cjw6"] Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.573404 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.581006 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-6cjw6"] Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.732439 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f0980bc5-5ad8-4479-ab3a-ea9c700428bd-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-6cjw6\" (UID: \"f0980bc5-5ad8-4479-ab3a-ea9c700428bd\") " pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.732547 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqnl4\" (UniqueName: \"kubernetes.io/projected/f0980bc5-5ad8-4479-ab3a-ea9c700428bd-kube-api-access-dqnl4\") pod \"dnsmasq-dns-5ccc8479f9-6cjw6\" (UID: \"f0980bc5-5ad8-4479-ab3a-ea9c700428bd\") " pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.732602 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0980bc5-5ad8-4479-ab3a-ea9c700428bd-config\") pod \"dnsmasq-dns-5ccc8479f9-6cjw6\" (UID: \"f0980bc5-5ad8-4479-ab3a-ea9c700428bd\") " pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.821581 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z9lkn"] Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.833744 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f0980bc5-5ad8-4479-ab3a-ea9c700428bd-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-6cjw6\" (UID: \"f0980bc5-5ad8-4479-ab3a-ea9c700428bd\") " pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.833832 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqnl4\" (UniqueName: \"kubernetes.io/projected/f0980bc5-5ad8-4479-ab3a-ea9c700428bd-kube-api-access-dqnl4\") pod \"dnsmasq-dns-5ccc8479f9-6cjw6\" (UID: \"f0980bc5-5ad8-4479-ab3a-ea9c700428bd\") " pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.833894 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0980bc5-5ad8-4479-ab3a-ea9c700428bd-config\") pod \"dnsmasq-dns-5ccc8479f9-6cjw6\" (UID: \"f0980bc5-5ad8-4479-ab3a-ea9c700428bd\") " pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.835267 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f0980bc5-5ad8-4479-ab3a-ea9c700428bd-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-6cjw6\" (UID: \"f0980bc5-5ad8-4479-ab3a-ea9c700428bd\") " pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.835385 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0980bc5-5ad8-4479-ab3a-ea9c700428bd-config\") pod \"dnsmasq-dns-5ccc8479f9-6cjw6\" (UID: \"f0980bc5-5ad8-4479-ab3a-ea9c700428bd\") " pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.846443 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ksnd5"] Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.850368 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.867303 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqnl4\" (UniqueName: \"kubernetes.io/projected/f0980bc5-5ad8-4479-ab3a-ea9c700428bd-kube-api-access-dqnl4\") pod \"dnsmasq-dns-5ccc8479f9-6cjw6\" (UID: \"f0980bc5-5ad8-4479-ab3a-ea9c700428bd\") " pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.894271 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ksnd5"] Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.896841 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.938466 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc31e29a-535b-401c-bc91-b146d3a53617-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-ksnd5\" (UID: \"fc31e29a-535b-401c-bc91-b146d3a53617\") " pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.938547 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlc8m\" (UniqueName: \"kubernetes.io/projected/fc31e29a-535b-401c-bc91-b146d3a53617-kube-api-access-hlc8m\") pod \"dnsmasq-dns-57d769cc4f-ksnd5\" (UID: \"fc31e29a-535b-401c-bc91-b146d3a53617\") " pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" Mar 17 09:29:14 crc kubenswrapper[4813]: I0317 09:29:14.938584 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc31e29a-535b-401c-bc91-b146d3a53617-config\") pod \"dnsmasq-dns-57d769cc4f-ksnd5\" (UID: \"fc31e29a-535b-401c-bc91-b146d3a53617\") " pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.040436 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlc8m\" (UniqueName: \"kubernetes.io/projected/fc31e29a-535b-401c-bc91-b146d3a53617-kube-api-access-hlc8m\") pod \"dnsmasq-dns-57d769cc4f-ksnd5\" (UID: \"fc31e29a-535b-401c-bc91-b146d3a53617\") " pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.040742 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc31e29a-535b-401c-bc91-b146d3a53617-config\") pod \"dnsmasq-dns-57d769cc4f-ksnd5\" (UID: \"fc31e29a-535b-401c-bc91-b146d3a53617\") " pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.040807 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc31e29a-535b-401c-bc91-b146d3a53617-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-ksnd5\" (UID: \"fc31e29a-535b-401c-bc91-b146d3a53617\") " pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.041884 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc31e29a-535b-401c-bc91-b146d3a53617-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-ksnd5\" (UID: \"fc31e29a-535b-401c-bc91-b146d3a53617\") " pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.042516 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc31e29a-535b-401c-bc91-b146d3a53617-config\") pod \"dnsmasq-dns-57d769cc4f-ksnd5\" (UID: \"fc31e29a-535b-401c-bc91-b146d3a53617\") " pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.059439 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlc8m\" (UniqueName: \"kubernetes.io/projected/fc31e29a-535b-401c-bc91-b146d3a53617-kube-api-access-hlc8m\") pod \"dnsmasq-dns-57d769cc4f-ksnd5\" (UID: \"fc31e29a-535b-401c-bc91-b146d3a53617\") " pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.193532 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.374029 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-6cjw6"] Mar 17 09:29:15 crc kubenswrapper[4813]: W0317 09:29:15.388355 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0980bc5_5ad8_4479_ab3a_ea9c700428bd.slice/crio-0b89a1845b8f7ce226b13811970de90ee876c1ec9864ed40cdfa798aba66db91 WatchSource:0}: Error finding container 0b89a1845b8f7ce226b13811970de90ee876c1ec9864ed40cdfa798aba66db91: Status 404 returned error can't find the container with id 0b89a1845b8f7ce226b13811970de90ee876c1ec9864ed40cdfa798aba66db91 Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.613188 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ksnd5"] Mar 17 09:29:15 crc kubenswrapper[4813]: W0317 09:29:15.622227 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfc31e29a_535b_401c_bc91_b146d3a53617.slice/crio-669e972f9a41e2309b46825715b397c68337c9f6a866fd690447fdfb95bb1b32 WatchSource:0}: Error finding container 669e972f9a41e2309b46825715b397c68337c9f6a866fd690447fdfb95bb1b32: Status 404 returned error can't find the container with id 669e972f9a41e2309b46825715b397c68337c9f6a866fd690447fdfb95bb1b32 Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.715730 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.720372 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.723517 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.723558 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.723779 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.723920 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.723948 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.724079 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-5rdjh" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.723928 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.725559 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.853504 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.853569 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.853619 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.853662 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m28ms\" (UniqueName: \"kubernetes.io/projected/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-kube-api-access-m28ms\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.853705 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.853740 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.853758 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.853774 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.853790 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.853815 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.853828 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.955830 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.955884 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.955905 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.955941 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m28ms\" (UniqueName: \"kubernetes.io/projected/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-kube-api-access-m28ms\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.955979 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.956012 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.956031 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.956050 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.956067 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.956097 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.956112 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.956522 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.959126 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.959687 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.959953 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.960250 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.960448 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.986289 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.987898 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.991777 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m28ms\" (UniqueName: \"kubernetes.io/projected/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-kube-api-access-m28ms\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:15 crc kubenswrapper[4813]: I0317 09:29:15.993164 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.000775 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.002903 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.003495 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.006459 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.006462 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.006742 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.006865 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.007199 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-xfkfx" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.007375 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.008553 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.014644 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.025886 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.042907 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" event={"ID":"f0980bc5-5ad8-4479-ab3a-ea9c700428bd","Type":"ContainerStarted","Data":"0b89a1845b8f7ce226b13811970de90ee876c1ec9864ed40cdfa798aba66db91"} Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.044185 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.049159 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" event={"ID":"fc31e29a-535b-401c-bc91-b146d3a53617","Type":"ContainerStarted","Data":"669e972f9a41e2309b46825715b397c68337c9f6a866fd690447fdfb95bb1b32"} Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.159921 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.159968 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.159994 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2517b76a-309a-4c9e-88e7-6061d528acc8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.160029 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2517b76a-309a-4c9e-88e7-6061d528acc8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.160044 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2517b76a-309a-4c9e-88e7-6061d528acc8-config-data\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.160058 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.160103 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcnpw\" (UniqueName: \"kubernetes.io/projected/2517b76a-309a-4c9e-88e7-6061d528acc8-kube-api-access-rcnpw\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.160120 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2517b76a-309a-4c9e-88e7-6061d528acc8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.160137 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.160163 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.160189 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2517b76a-309a-4c9e-88e7-6061d528acc8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.261233 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2517b76a-309a-4c9e-88e7-6061d528acc8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.261275 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2517b76a-309a-4c9e-88e7-6061d528acc8-config-data\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.261293 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.261341 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcnpw\" (UniqueName: \"kubernetes.io/projected/2517b76a-309a-4c9e-88e7-6061d528acc8-kube-api-access-rcnpw\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.261358 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2517b76a-309a-4c9e-88e7-6061d528acc8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.261379 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.261408 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.261431 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2517b76a-309a-4c9e-88e7-6061d528acc8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.261455 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.261473 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.261490 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2517b76a-309a-4c9e-88e7-6061d528acc8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.262468 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.263055 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2517b76a-309a-4c9e-88e7-6061d528acc8-config-data\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.263260 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2517b76a-309a-4c9e-88e7-6061d528acc8-server-conf\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.263712 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2517b76a-309a-4c9e-88e7-6061d528acc8-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.263920 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.264004 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.266220 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.266400 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2517b76a-309a-4c9e-88e7-6061d528acc8-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.266510 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2517b76a-309a-4c9e-88e7-6061d528acc8-pod-info\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.275868 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.278201 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcnpw\" (UniqueName: \"kubernetes.io/projected/2517b76a-309a-4c9e-88e7-6061d528acc8-kube-api-access-rcnpw\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.282861 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " pod="openstack/rabbitmq-server-0" Mar 17 09:29:16 crc kubenswrapper[4813]: I0317 09:29:16.363629 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.065776 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.067682 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.070211 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-sqpr9" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.071053 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.071192 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.073002 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.080044 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.082390 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.176376 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.176485 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9ad18f78-70c7-4b1d-bda7-ed338fdcdd64-config-data-default\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.176523 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbpdl\" (UniqueName: \"kubernetes.io/projected/9ad18f78-70c7-4b1d-bda7-ed338fdcdd64-kube-api-access-hbpdl\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.176558 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ad18f78-70c7-4b1d-bda7-ed338fdcdd64-operator-scripts\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.176629 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ad18f78-70c7-4b1d-bda7-ed338fdcdd64-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.176686 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9ad18f78-70c7-4b1d-bda7-ed338fdcdd64-config-data-generated\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.176726 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ad18f78-70c7-4b1d-bda7-ed338fdcdd64-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.177419 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9ad18f78-70c7-4b1d-bda7-ed338fdcdd64-kolla-config\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.279123 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.279219 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9ad18f78-70c7-4b1d-bda7-ed338fdcdd64-config-data-default\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.279252 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbpdl\" (UniqueName: \"kubernetes.io/projected/9ad18f78-70c7-4b1d-bda7-ed338fdcdd64-kube-api-access-hbpdl\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.279295 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ad18f78-70c7-4b1d-bda7-ed338fdcdd64-operator-scripts\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.279330 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ad18f78-70c7-4b1d-bda7-ed338fdcdd64-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.279385 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9ad18f78-70c7-4b1d-bda7-ed338fdcdd64-config-data-generated\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.279425 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ad18f78-70c7-4b1d-bda7-ed338fdcdd64-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.279487 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.279520 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9ad18f78-70c7-4b1d-bda7-ed338fdcdd64-kolla-config\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.280740 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9ad18f78-70c7-4b1d-bda7-ed338fdcdd64-kolla-config\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.281039 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/9ad18f78-70c7-4b1d-bda7-ed338fdcdd64-config-data-generated\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.281157 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9ad18f78-70c7-4b1d-bda7-ed338fdcdd64-operator-scripts\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.281961 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/9ad18f78-70c7-4b1d-bda7-ed338fdcdd64-config-data-default\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.285331 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ad18f78-70c7-4b1d-bda7-ed338fdcdd64-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.285803 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ad18f78-70c7-4b1d-bda7-ed338fdcdd64-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.304199 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbpdl\" (UniqueName: \"kubernetes.io/projected/9ad18f78-70c7-4b1d-bda7-ed338fdcdd64-kube-api-access-hbpdl\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.322974 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-0\" (UID: \"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64\") " pod="openstack/openstack-galera-0" Mar 17 09:29:17 crc kubenswrapper[4813]: I0317 09:29:17.389442 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.552985 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.554691 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.558157 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.558360 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.558551 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.558723 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-25kcl" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.571549 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.701067 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.701128 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.701158 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.701178 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.701239 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.701276 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.701303 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.701334 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wl76r\" (UniqueName: \"kubernetes.io/projected/95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1-kube-api-access-wl76r\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.803563 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.803667 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.803697 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.803749 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.803804 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.803844 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.803891 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wl76r\" (UniqueName: \"kubernetes.io/projected/95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1-kube-api-access-wl76r\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.803924 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.803995 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.804788 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.804819 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.804872 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.805179 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.814733 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.819244 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.820147 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wl76r\" (UniqueName: \"kubernetes.io/projected/95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1-kube-api-access-wl76r\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.836382 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-cell1-galera-0\" (UID: \"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1\") " pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.884371 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.897153 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.898028 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.903187 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.903418 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-v76hk" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.904296 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Mar 17 09:29:18 crc kubenswrapper[4813]: I0317 09:29:18.913258 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 17 09:29:19 crc kubenswrapper[4813]: I0317 09:29:19.005738 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rfhqx\" (UniqueName: \"kubernetes.io/projected/c43912b9-949e-4d64-ae70-e59594cc329a-kube-api-access-rfhqx\") pod \"memcached-0\" (UID: \"c43912b9-949e-4d64-ae70-e59594cc329a\") " pod="openstack/memcached-0" Mar 17 09:29:19 crc kubenswrapper[4813]: I0317 09:29:19.005803 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c43912b9-949e-4d64-ae70-e59594cc329a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c43912b9-949e-4d64-ae70-e59594cc329a\") " pod="openstack/memcached-0" Mar 17 09:29:19 crc kubenswrapper[4813]: I0317 09:29:19.005849 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c43912b9-949e-4d64-ae70-e59594cc329a-config-data\") pod \"memcached-0\" (UID: \"c43912b9-949e-4d64-ae70-e59594cc329a\") " pod="openstack/memcached-0" Mar 17 09:29:19 crc kubenswrapper[4813]: I0317 09:29:19.005875 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c43912b9-949e-4d64-ae70-e59594cc329a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c43912b9-949e-4d64-ae70-e59594cc329a\") " pod="openstack/memcached-0" Mar 17 09:29:19 crc kubenswrapper[4813]: I0317 09:29:19.006040 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c43912b9-949e-4d64-ae70-e59594cc329a-kolla-config\") pod \"memcached-0\" (UID: \"c43912b9-949e-4d64-ae70-e59594cc329a\") " pod="openstack/memcached-0" Mar 17 09:29:19 crc kubenswrapper[4813]: I0317 09:29:19.107606 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c43912b9-949e-4d64-ae70-e59594cc329a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c43912b9-949e-4d64-ae70-e59594cc329a\") " pod="openstack/memcached-0" Mar 17 09:29:19 crc kubenswrapper[4813]: I0317 09:29:19.107691 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c43912b9-949e-4d64-ae70-e59594cc329a-config-data\") pod \"memcached-0\" (UID: \"c43912b9-949e-4d64-ae70-e59594cc329a\") " pod="openstack/memcached-0" Mar 17 09:29:19 crc kubenswrapper[4813]: I0317 09:29:19.107722 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c43912b9-949e-4d64-ae70-e59594cc329a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c43912b9-949e-4d64-ae70-e59594cc329a\") " pod="openstack/memcached-0" Mar 17 09:29:19 crc kubenswrapper[4813]: I0317 09:29:19.107772 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c43912b9-949e-4d64-ae70-e59594cc329a-kolla-config\") pod \"memcached-0\" (UID: \"c43912b9-949e-4d64-ae70-e59594cc329a\") " pod="openstack/memcached-0" Mar 17 09:29:19 crc kubenswrapper[4813]: I0317 09:29:19.107794 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rfhqx\" (UniqueName: \"kubernetes.io/projected/c43912b9-949e-4d64-ae70-e59594cc329a-kube-api-access-rfhqx\") pod \"memcached-0\" (UID: \"c43912b9-949e-4d64-ae70-e59594cc329a\") " pod="openstack/memcached-0" Mar 17 09:29:19 crc kubenswrapper[4813]: I0317 09:29:19.108816 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c43912b9-949e-4d64-ae70-e59594cc329a-config-data\") pod \"memcached-0\" (UID: \"c43912b9-949e-4d64-ae70-e59594cc329a\") " pod="openstack/memcached-0" Mar 17 09:29:19 crc kubenswrapper[4813]: I0317 09:29:19.108844 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/c43912b9-949e-4d64-ae70-e59594cc329a-kolla-config\") pod \"memcached-0\" (UID: \"c43912b9-949e-4d64-ae70-e59594cc329a\") " pod="openstack/memcached-0" Mar 17 09:29:19 crc kubenswrapper[4813]: I0317 09:29:19.112105 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c43912b9-949e-4d64-ae70-e59594cc329a-combined-ca-bundle\") pod \"memcached-0\" (UID: \"c43912b9-949e-4d64-ae70-e59594cc329a\") " pod="openstack/memcached-0" Mar 17 09:29:19 crc kubenswrapper[4813]: I0317 09:29:19.112127 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/c43912b9-949e-4d64-ae70-e59594cc329a-memcached-tls-certs\") pod \"memcached-0\" (UID: \"c43912b9-949e-4d64-ae70-e59594cc329a\") " pod="openstack/memcached-0" Mar 17 09:29:19 crc kubenswrapper[4813]: I0317 09:29:19.123203 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rfhqx\" (UniqueName: \"kubernetes.io/projected/c43912b9-949e-4d64-ae70-e59594cc329a-kube-api-access-rfhqx\") pod \"memcached-0\" (UID: \"c43912b9-949e-4d64-ae70-e59594cc329a\") " pod="openstack/memcached-0" Mar 17 09:29:19 crc kubenswrapper[4813]: I0317 09:29:19.215589 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Mar 17 09:29:20 crc kubenswrapper[4813]: I0317 09:29:20.964102 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 17 09:29:20 crc kubenswrapper[4813]: I0317 09:29:20.966198 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 17 09:29:20 crc kubenswrapper[4813]: I0317 09:29:20.968718 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-jmmzv" Mar 17 09:29:20 crc kubenswrapper[4813]: I0317 09:29:20.981849 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 17 09:29:21 crc kubenswrapper[4813]: I0317 09:29:21.047707 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcccg\" (UniqueName: \"kubernetes.io/projected/59209cd3-5db8-4ddc-806e-1db457164c1f-kube-api-access-dcccg\") pod \"kube-state-metrics-0\" (UID: \"59209cd3-5db8-4ddc-806e-1db457164c1f\") " pod="openstack/kube-state-metrics-0" Mar 17 09:29:21 crc kubenswrapper[4813]: I0317 09:29:21.149239 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcccg\" (UniqueName: \"kubernetes.io/projected/59209cd3-5db8-4ddc-806e-1db457164c1f-kube-api-access-dcccg\") pod \"kube-state-metrics-0\" (UID: \"59209cd3-5db8-4ddc-806e-1db457164c1f\") " pod="openstack/kube-state-metrics-0" Mar 17 09:29:21 crc kubenswrapper[4813]: I0317 09:29:21.166379 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcccg\" (UniqueName: \"kubernetes.io/projected/59209cd3-5db8-4ddc-806e-1db457164c1f-kube-api-access-dcccg\") pod \"kube-state-metrics-0\" (UID: \"59209cd3-5db8-4ddc-806e-1db457164c1f\") " pod="openstack/kube-state-metrics-0" Mar 17 09:29:21 crc kubenswrapper[4813]: I0317 09:29:21.288950 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.254033 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vdc6c"] Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.255596 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.257010 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.259969 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-zzcvs" Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.260289 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.268676 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vdc6c"] Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.323542 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-bt47r"] Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.325822 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.333965 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-bt47r"] Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.401314 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3752b62f-49dd-46a7-ae34-9b6b20343a09-etc-ovs\") pod \"ovn-controller-ovs-bt47r\" (UID: \"3752b62f-49dd-46a7-ae34-9b6b20343a09\") " pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.401682 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4090993b-51ce-4ce3-a6d6-a1501ab3ba05-var-run-ovn\") pod \"ovn-controller-vdc6c\" (UID: \"4090993b-51ce-4ce3-a6d6-a1501ab3ba05\") " pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.401758 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3752b62f-49dd-46a7-ae34-9b6b20343a09-var-lib\") pod \"ovn-controller-ovs-bt47r\" (UID: \"3752b62f-49dd-46a7-ae34-9b6b20343a09\") " pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.401778 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znwmt\" (UniqueName: \"kubernetes.io/projected/3752b62f-49dd-46a7-ae34-9b6b20343a09-kube-api-access-znwmt\") pod \"ovn-controller-ovs-bt47r\" (UID: \"3752b62f-49dd-46a7-ae34-9b6b20343a09\") " pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.401797 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3752b62f-49dd-46a7-ae34-9b6b20343a09-var-log\") pod \"ovn-controller-ovs-bt47r\" (UID: \"3752b62f-49dd-46a7-ae34-9b6b20343a09\") " pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.401824 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4090993b-51ce-4ce3-a6d6-a1501ab3ba05-scripts\") pod \"ovn-controller-vdc6c\" (UID: \"4090993b-51ce-4ce3-a6d6-a1501ab3ba05\") " pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.401916 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3752b62f-49dd-46a7-ae34-9b6b20343a09-scripts\") pod \"ovn-controller-ovs-bt47r\" (UID: \"3752b62f-49dd-46a7-ae34-9b6b20343a09\") " pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.401941 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9cvj\" (UniqueName: \"kubernetes.io/projected/4090993b-51ce-4ce3-a6d6-a1501ab3ba05-kube-api-access-g9cvj\") pod \"ovn-controller-vdc6c\" (UID: \"4090993b-51ce-4ce3-a6d6-a1501ab3ba05\") " pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.401967 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4090993b-51ce-4ce3-a6d6-a1501ab3ba05-combined-ca-bundle\") pod \"ovn-controller-vdc6c\" (UID: \"4090993b-51ce-4ce3-a6d6-a1501ab3ba05\") " pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.402000 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4090993b-51ce-4ce3-a6d6-a1501ab3ba05-var-log-ovn\") pod \"ovn-controller-vdc6c\" (UID: \"4090993b-51ce-4ce3-a6d6-a1501ab3ba05\") " pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.402165 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/4090993b-51ce-4ce3-a6d6-a1501ab3ba05-ovn-controller-tls-certs\") pod \"ovn-controller-vdc6c\" (UID: \"4090993b-51ce-4ce3-a6d6-a1501ab3ba05\") " pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.402187 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3752b62f-49dd-46a7-ae34-9b6b20343a09-var-run\") pod \"ovn-controller-ovs-bt47r\" (UID: \"3752b62f-49dd-46a7-ae34-9b6b20343a09\") " pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.402210 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4090993b-51ce-4ce3-a6d6-a1501ab3ba05-var-run\") pod \"ovn-controller-vdc6c\" (UID: \"4090993b-51ce-4ce3-a6d6-a1501ab3ba05\") " pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.503572 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3752b62f-49dd-46a7-ae34-9b6b20343a09-var-lib\") pod \"ovn-controller-ovs-bt47r\" (UID: \"3752b62f-49dd-46a7-ae34-9b6b20343a09\") " pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.503632 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znwmt\" (UniqueName: \"kubernetes.io/projected/3752b62f-49dd-46a7-ae34-9b6b20343a09-kube-api-access-znwmt\") pod \"ovn-controller-ovs-bt47r\" (UID: \"3752b62f-49dd-46a7-ae34-9b6b20343a09\") " pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:24 crc kubenswrapper[4813]: I0317 09:29:24.503650 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3752b62f-49dd-46a7-ae34-9b6b20343a09-var-log\") pod \"ovn-controller-ovs-bt47r\" (UID: \"3752b62f-49dd-46a7-ae34-9b6b20343a09\") " pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.503674 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4090993b-51ce-4ce3-a6d6-a1501ab3ba05-scripts\") pod \"ovn-controller-vdc6c\" (UID: \"4090993b-51ce-4ce3-a6d6-a1501ab3ba05\") " pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.503702 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3752b62f-49dd-46a7-ae34-9b6b20343a09-scripts\") pod \"ovn-controller-ovs-bt47r\" (UID: \"3752b62f-49dd-46a7-ae34-9b6b20343a09\") " pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.503722 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9cvj\" (UniqueName: \"kubernetes.io/projected/4090993b-51ce-4ce3-a6d6-a1501ab3ba05-kube-api-access-g9cvj\") pod \"ovn-controller-vdc6c\" (UID: \"4090993b-51ce-4ce3-a6d6-a1501ab3ba05\") " pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.503737 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4090993b-51ce-4ce3-a6d6-a1501ab3ba05-combined-ca-bundle\") pod \"ovn-controller-vdc6c\" (UID: \"4090993b-51ce-4ce3-a6d6-a1501ab3ba05\") " pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.503756 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4090993b-51ce-4ce3-a6d6-a1501ab3ba05-var-log-ovn\") pod \"ovn-controller-vdc6c\" (UID: \"4090993b-51ce-4ce3-a6d6-a1501ab3ba05\") " pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.503820 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/4090993b-51ce-4ce3-a6d6-a1501ab3ba05-ovn-controller-tls-certs\") pod \"ovn-controller-vdc6c\" (UID: \"4090993b-51ce-4ce3-a6d6-a1501ab3ba05\") " pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.503837 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3752b62f-49dd-46a7-ae34-9b6b20343a09-var-run\") pod \"ovn-controller-ovs-bt47r\" (UID: \"3752b62f-49dd-46a7-ae34-9b6b20343a09\") " pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.503853 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4090993b-51ce-4ce3-a6d6-a1501ab3ba05-var-run\") pod \"ovn-controller-vdc6c\" (UID: \"4090993b-51ce-4ce3-a6d6-a1501ab3ba05\") " pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.503871 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3752b62f-49dd-46a7-ae34-9b6b20343a09-etc-ovs\") pod \"ovn-controller-ovs-bt47r\" (UID: \"3752b62f-49dd-46a7-ae34-9b6b20343a09\") " pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.503888 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4090993b-51ce-4ce3-a6d6-a1501ab3ba05-var-run-ovn\") pod \"ovn-controller-vdc6c\" (UID: \"4090993b-51ce-4ce3-a6d6-a1501ab3ba05\") " pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.504453 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/3752b62f-49dd-46a7-ae34-9b6b20343a09-var-lib\") pod \"ovn-controller-ovs-bt47r\" (UID: \"3752b62f-49dd-46a7-ae34-9b6b20343a09\") " pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.504520 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4090993b-51ce-4ce3-a6d6-a1501ab3ba05-var-run-ovn\") pod \"ovn-controller-vdc6c\" (UID: \"4090993b-51ce-4ce3-a6d6-a1501ab3ba05\") " pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.504568 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/3752b62f-49dd-46a7-ae34-9b6b20343a09-var-log\") pod \"ovn-controller-ovs-bt47r\" (UID: \"3752b62f-49dd-46a7-ae34-9b6b20343a09\") " pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.504643 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/3752b62f-49dd-46a7-ae34-9b6b20343a09-var-run\") pod \"ovn-controller-ovs-bt47r\" (UID: \"3752b62f-49dd-46a7-ae34-9b6b20343a09\") " pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.504713 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4090993b-51ce-4ce3-a6d6-a1501ab3ba05-var-run\") pod \"ovn-controller-vdc6c\" (UID: \"4090993b-51ce-4ce3-a6d6-a1501ab3ba05\") " pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.504841 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/3752b62f-49dd-46a7-ae34-9b6b20343a09-etc-ovs\") pod \"ovn-controller-ovs-bt47r\" (UID: \"3752b62f-49dd-46a7-ae34-9b6b20343a09\") " pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.504926 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4090993b-51ce-4ce3-a6d6-a1501ab3ba05-var-log-ovn\") pod \"ovn-controller-vdc6c\" (UID: \"4090993b-51ce-4ce3-a6d6-a1501ab3ba05\") " pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.508118 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4090993b-51ce-4ce3-a6d6-a1501ab3ba05-scripts\") pod \"ovn-controller-vdc6c\" (UID: \"4090993b-51ce-4ce3-a6d6-a1501ab3ba05\") " pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.509126 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3752b62f-49dd-46a7-ae34-9b6b20343a09-scripts\") pod \"ovn-controller-ovs-bt47r\" (UID: \"3752b62f-49dd-46a7-ae34-9b6b20343a09\") " pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.520917 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4090993b-51ce-4ce3-a6d6-a1501ab3ba05-combined-ca-bundle\") pod \"ovn-controller-vdc6c\" (UID: \"4090993b-51ce-4ce3-a6d6-a1501ab3ba05\") " pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.526046 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9cvj\" (UniqueName: \"kubernetes.io/projected/4090993b-51ce-4ce3-a6d6-a1501ab3ba05-kube-api-access-g9cvj\") pod \"ovn-controller-vdc6c\" (UID: \"4090993b-51ce-4ce3-a6d6-a1501ab3ba05\") " pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.526484 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znwmt\" (UniqueName: \"kubernetes.io/projected/3752b62f-49dd-46a7-ae34-9b6b20343a09-kube-api-access-znwmt\") pod \"ovn-controller-ovs-bt47r\" (UID: \"3752b62f-49dd-46a7-ae34-9b6b20343a09\") " pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.530035 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/4090993b-51ce-4ce3-a6d6-a1501ab3ba05-ovn-controller-tls-certs\") pod \"ovn-controller-vdc6c\" (UID: \"4090993b-51ce-4ce3-a6d6-a1501ab3ba05\") " pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.573666 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.651459 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.861924 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.865059 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.868134 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-tqcx2" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.868399 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.868756 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.869405 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.870406 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:24.874570 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.010757 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34e286fb-7dd1-4e85-89e3-e926b232f5a6-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.010827 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34e286fb-7dd1-4e85-89e3-e926b232f5a6-config\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.010907 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34e286fb-7dd1-4e85-89e3-e926b232f5a6-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.010966 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/34e286fb-7dd1-4e85-89e3-e926b232f5a6-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.011041 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/34e286fb-7dd1-4e85-89e3-e926b232f5a6-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.011072 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/34e286fb-7dd1-4e85-89e3-e926b232f5a6-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.011119 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkpzb\" (UniqueName: \"kubernetes.io/projected/34e286fb-7dd1-4e85-89e3-e926b232f5a6-kube-api-access-wkpzb\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.011147 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.112542 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34e286fb-7dd1-4e85-89e3-e926b232f5a6-config\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.112656 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34e286fb-7dd1-4e85-89e3-e926b232f5a6-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.112695 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/34e286fb-7dd1-4e85-89e3-e926b232f5a6-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.112728 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/34e286fb-7dd1-4e85-89e3-e926b232f5a6-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.112751 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/34e286fb-7dd1-4e85-89e3-e926b232f5a6-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.112789 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkpzb\" (UniqueName: \"kubernetes.io/projected/34e286fb-7dd1-4e85-89e3-e926b232f5a6-kube-api-access-wkpzb\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.112818 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.112872 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34e286fb-7dd1-4e85-89e3-e926b232f5a6-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.113926 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.114117 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/34e286fb-7dd1-4e85-89e3-e926b232f5a6-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.117579 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/34e286fb-7dd1-4e85-89e3-e926b232f5a6-config\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.117655 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/34e286fb-7dd1-4e85-89e3-e926b232f5a6-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.118682 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/34e286fb-7dd1-4e85-89e3-e926b232f5a6-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.121214 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/34e286fb-7dd1-4e85-89e3-e926b232f5a6-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.122834 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/34e286fb-7dd1-4e85-89e3-e926b232f5a6-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.133110 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkpzb\" (UniqueName: \"kubernetes.io/projected/34e286fb-7dd1-4e85-89e3-e926b232f5a6-kube-api-access-wkpzb\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.135931 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"ovsdbserver-nb-0\" (UID: \"34e286fb-7dd1-4e85-89e3-e926b232f5a6\") " pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:25 crc kubenswrapper[4813]: I0317 09:29:25.203415 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:27 crc kubenswrapper[4813]: E0317 09:29:27.348222 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 17 09:29:27 crc kubenswrapper[4813]: E0317 09:29:27.349785 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lzgpg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-988sz_openstack(63458ae3-263a-4d68-95bd-4824acb4cebd): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 09:29:27 crc kubenswrapper[4813]: E0317 09:29:27.350996 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-988sz" podUID="63458ae3-263a-4d68-95bd-4824acb4cebd" Mar 17 09:29:27 crc kubenswrapper[4813]: E0317 09:29:27.359037 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Mar 17 09:29:27 crc kubenswrapper[4813]: E0317 09:29:27.359360 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-sqrfw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-z9lkn_openstack(8369f72b-14fd-40f4-864c-c8bddd0f43e8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 09:29:27 crc kubenswrapper[4813]: E0317 09:29:27.360538 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-z9lkn" podUID="8369f72b-14fd-40f4-864c-c8bddd0f43e8" Mar 17 09:29:27 crc kubenswrapper[4813]: I0317 09:29:27.782811 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 17 09:29:27 crc kubenswrapper[4813]: W0317 09:29:27.820288 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2517b76a_309a_4c9e_88e7_6061d528acc8.slice/crio-f07311e221541fb35044925dc0c83c28c25c05e7d66b76a0cb68aef674ef228f WatchSource:0}: Error finding container f07311e221541fb35044925dc0c83c28c25c05e7d66b76a0cb68aef674ef228f: Status 404 returned error can't find the container with id f07311e221541fb35044925dc0c83c28c25c05e7d66b76a0cb68aef674ef228f Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.088030 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.089158 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.092388 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.095950 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-lstkk" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.095965 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.096146 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.130820 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.147412 4813 generic.go:334] "Generic (PLEG): container finished" podID="fc31e29a-535b-401c-bc91-b146d3a53617" containerID="14dbda82b6001173ff09d165849f6a30a02b46b5c74614a824fb80ca14c07ff5" exitCode=0 Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.147486 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" event={"ID":"fc31e29a-535b-401c-bc91-b146d3a53617","Type":"ContainerDied","Data":"14dbda82b6001173ff09d165849f6a30a02b46b5c74614a824fb80ca14c07ff5"} Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.149174 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2517b76a-309a-4c9e-88e7-6061d528acc8","Type":"ContainerStarted","Data":"f07311e221541fb35044925dc0c83c28c25c05e7d66b76a0cb68aef674ef228f"} Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.151253 4813 generic.go:334] "Generic (PLEG): container finished" podID="f0980bc5-5ad8-4479-ab3a-ea9c700428bd" containerID="0c490e030ba62559a80efe6fad223b7e79fd91130b84beb558b2252202881d57" exitCode=0 Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.151387 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" event={"ID":"f0980bc5-5ad8-4479-ab3a-ea9c700428bd","Type":"ContainerDied","Data":"0c490e030ba62559a80efe6fad223b7e79fd91130b84beb558b2252202881d57"} Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.190414 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7e4c8a33-829e-4aec-be7c-0dad92ce3916-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.190499 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e4c8a33-829e-4aec-be7c-0dad92ce3916-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.190553 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e4c8a33-829e-4aec-be7c-0dad92ce3916-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.190590 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e4c8a33-829e-4aec-be7c-0dad92ce3916-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.190647 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e4c8a33-829e-4aec-be7c-0dad92ce3916-config\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.190694 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2vc7\" (UniqueName: \"kubernetes.io/projected/7e4c8a33-829e-4aec-be7c-0dad92ce3916-kube-api-access-z2vc7\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.190743 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e4c8a33-829e-4aec-be7c-0dad92ce3916-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.190815 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.255132 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.275215 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vdc6c"] Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.287638 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.292875 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e4c8a33-829e-4aec-be7c-0dad92ce3916-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.292951 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e4c8a33-829e-4aec-be7c-0dad92ce3916-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.292991 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e4c8a33-829e-4aec-be7c-0dad92ce3916-config\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.293038 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2vc7\" (UniqueName: \"kubernetes.io/projected/7e4c8a33-829e-4aec-be7c-0dad92ce3916-kube-api-access-z2vc7\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.293091 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e4c8a33-829e-4aec-be7c-0dad92ce3916-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.293132 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.293550 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7e4c8a33-829e-4aec-be7c-0dad92ce3916-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.293637 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e4c8a33-829e-4aec-be7c-0dad92ce3916-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.294078 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e4c8a33-829e-4aec-be7c-0dad92ce3916-config\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.297515 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.298709 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.299533 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/7e4c8a33-829e-4aec-be7c-0dad92ce3916-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.299935 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/7e4c8a33-829e-4aec-be7c-0dad92ce3916-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.300210 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e4c8a33-829e-4aec-be7c-0dad92ce3916-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.303713 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7e4c8a33-829e-4aec-be7c-0dad92ce3916-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.306389 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7e4c8a33-829e-4aec-be7c-0dad92ce3916-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.309691 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.319386 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2vc7\" (UniqueName: \"kubernetes.io/projected/7e4c8a33-829e-4aec-be7c-0dad92ce3916-kube-api-access-z2vc7\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.323334 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.332725 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"7e4c8a33-829e-4aec-be7c-0dad92ce3916\") " pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.385187 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-bt47r"] Mar 17 09:29:28 crc kubenswrapper[4813]: W0317 09:29:28.388218 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3752b62f_49dd_46a7_ae34_9b6b20343a09.slice/crio-558c7ae1a2d2494d0a5752578c732c9e6a86618a8b63e5078ba223bedb9ce09e WatchSource:0}: Error finding container 558c7ae1a2d2494d0a5752578c732c9e6a86618a8b63e5078ba223bedb9ce09e: Status 404 returned error can't find the container with id 558c7ae1a2d2494d0a5752578c732c9e6a86618a8b63e5078ba223bedb9ce09e Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.414453 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.524220 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Mar 17 09:29:28 crc kubenswrapper[4813]: W0317 09:29:28.527116 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod34e286fb_7dd1_4e85_89e3_e926b232f5a6.slice/crio-44c57ad69b2a36c5b73ce2dd789f275473dfb0673753aab2eedfaf0d83e4aac9 WatchSource:0}: Error finding container 44c57ad69b2a36c5b73ce2dd789f275473dfb0673753aab2eedfaf0d83e4aac9: Status 404 returned error can't find the container with id 44c57ad69b2a36c5b73ce2dd789f275473dfb0673753aab2eedfaf0d83e4aac9 Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.630692 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-z9lkn" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.671995 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-988sz" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.699198 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8369f72b-14fd-40f4-864c-c8bddd0f43e8-dns-svc\") pod \"8369f72b-14fd-40f4-864c-c8bddd0f43e8\" (UID: \"8369f72b-14fd-40f4-864c-c8bddd0f43e8\") " Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.699573 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8369f72b-14fd-40f4-864c-c8bddd0f43e8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8369f72b-14fd-40f4-864c-c8bddd0f43e8" (UID: "8369f72b-14fd-40f4-864c-c8bddd0f43e8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.699656 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8369f72b-14fd-40f4-864c-c8bddd0f43e8-config\") pod \"8369f72b-14fd-40f4-864c-c8bddd0f43e8\" (UID: \"8369f72b-14fd-40f4-864c-c8bddd0f43e8\") " Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.699685 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sqrfw\" (UniqueName: \"kubernetes.io/projected/8369f72b-14fd-40f4-864c-c8bddd0f43e8-kube-api-access-sqrfw\") pod \"8369f72b-14fd-40f4-864c-c8bddd0f43e8\" (UID: \"8369f72b-14fd-40f4-864c-c8bddd0f43e8\") " Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.700100 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8369f72b-14fd-40f4-864c-c8bddd0f43e8-config" (OuterVolumeSpecName: "config") pod "8369f72b-14fd-40f4-864c-c8bddd0f43e8" (UID: "8369f72b-14fd-40f4-864c-c8bddd0f43e8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.700663 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8369f72b-14fd-40f4-864c-c8bddd0f43e8-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.700701 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8369f72b-14fd-40f4-864c-c8bddd0f43e8-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.704910 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8369f72b-14fd-40f4-864c-c8bddd0f43e8-kube-api-access-sqrfw" (OuterVolumeSpecName: "kube-api-access-sqrfw") pod "8369f72b-14fd-40f4-864c-c8bddd0f43e8" (UID: "8369f72b-14fd-40f4-864c-c8bddd0f43e8"). InnerVolumeSpecName "kube-api-access-sqrfw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.806701 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63458ae3-263a-4d68-95bd-4824acb4cebd-config\") pod \"63458ae3-263a-4d68-95bd-4824acb4cebd\" (UID: \"63458ae3-263a-4d68-95bd-4824acb4cebd\") " Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.806834 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzgpg\" (UniqueName: \"kubernetes.io/projected/63458ae3-263a-4d68-95bd-4824acb4cebd-kube-api-access-lzgpg\") pod \"63458ae3-263a-4d68-95bd-4824acb4cebd\" (UID: \"63458ae3-263a-4d68-95bd-4824acb4cebd\") " Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.807121 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/63458ae3-263a-4d68-95bd-4824acb4cebd-config" (OuterVolumeSpecName: "config") pod "63458ae3-263a-4d68-95bd-4824acb4cebd" (UID: "63458ae3-263a-4d68-95bd-4824acb4cebd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.807157 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sqrfw\" (UniqueName: \"kubernetes.io/projected/8369f72b-14fd-40f4-864c-c8bddd0f43e8-kube-api-access-sqrfw\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.811694 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63458ae3-263a-4d68-95bd-4824acb4cebd-kube-api-access-lzgpg" (OuterVolumeSpecName: "kube-api-access-lzgpg") pod "63458ae3-263a-4d68-95bd-4824acb4cebd" (UID: "63458ae3-263a-4d68-95bd-4824acb4cebd"). InnerVolumeSpecName "kube-api-access-lzgpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.909552 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/63458ae3-263a-4d68-95bd-4824acb4cebd-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.909677 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzgpg\" (UniqueName: \"kubernetes.io/projected/63458ae3-263a-4d68-95bd-4824acb4cebd-kube-api-access-lzgpg\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:28 crc kubenswrapper[4813]: I0317 09:29:28.966447 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Mar 17 09:29:29 crc kubenswrapper[4813]: W0317 09:29:29.091793 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e4c8a33_829e_4aec_be7c_0dad92ce3916.slice/crio-82e34d694e5edf2ae4e9d2e9ff1a0946920faedf6bd1edf7548cb007bb77a19d WatchSource:0}: Error finding container 82e34d694e5edf2ae4e9d2e9ff1a0946920faedf6bd1edf7548cb007bb77a19d: Status 404 returned error can't find the container with id 82e34d694e5edf2ae4e9d2e9ff1a0946920faedf6bd1edf7548cb007bb77a19d Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.160903 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753","Type":"ContainerStarted","Data":"f419aa470ce9c02a701fd2d50b2ce9af2a29a9d337962090cd3f469c436d39af"} Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.162865 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"34e286fb-7dd1-4e85-89e3-e926b232f5a6","Type":"ContainerStarted","Data":"44c57ad69b2a36c5b73ce2dd789f275473dfb0673753aab2eedfaf0d83e4aac9"} Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.165134 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" event={"ID":"fc31e29a-535b-401c-bc91-b146d3a53617","Type":"ContainerStarted","Data":"062968e76b36dc557aa11d7a139d611432719617f05f9c248fc9754be1a25125"} Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.165269 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.166377 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64","Type":"ContainerStarted","Data":"20dfe6e6928a077a946f43e73c12b0f684ea6f54a4f07a5c58c2551578a3e3db"} Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.167712 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"59209cd3-5db8-4ddc-806e-1db457164c1f","Type":"ContainerStarted","Data":"ec5ec8ea069dc659eea9cb2219c465524921b7df81d24a13f4509fb998cd1c92"} Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.174046 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7e4c8a33-829e-4aec-be7c-0dad92ce3916","Type":"ContainerStarted","Data":"82e34d694e5edf2ae4e9d2e9ff1a0946920faedf6bd1edf7548cb007bb77a19d"} Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.175118 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vdc6c" event={"ID":"4090993b-51ce-4ce3-a6d6-a1501ab3ba05","Type":"ContainerStarted","Data":"b69441cb632e3696aa730392ee619c3ef8cfc1f6974218383edc26ee7279df9a"} Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.176619 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-988sz" event={"ID":"63458ae3-263a-4d68-95bd-4824acb4cebd","Type":"ContainerDied","Data":"a33ce5fe3932d6af3b62fc277032cd57b216275da6889e83b581d90b67bcfa9c"} Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.176725 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-988sz" Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.184777 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1","Type":"ContainerStarted","Data":"0ff3b70985154123a5f4485ca8f6fd7e582ca32b396a35495719b292cf396c28"} Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.187549 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c43912b9-949e-4d64-ae70-e59594cc329a","Type":"ContainerStarted","Data":"8b082fe0964de283054ed9c0d5da337de81bf7d2b758cd25eb7e5ed1351785ac"} Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.188997 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" podStartSLOduration=3.352167775 podStartE2EDuration="15.188973662s" podCreationTimestamp="2026-03-17 09:29:14 +0000 UTC" firstStartedPulling="2026-03-17 09:29:15.624910701 +0000 UTC m=+1177.725714200" lastFinishedPulling="2026-03-17 09:29:27.461716588 +0000 UTC m=+1189.562520087" observedRunningTime="2026-03-17 09:29:29.181796637 +0000 UTC m=+1191.282600136" watchObservedRunningTime="2026-03-17 09:29:29.188973662 +0000 UTC m=+1191.289777161" Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.190753 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" event={"ID":"f0980bc5-5ad8-4479-ab3a-ea9c700428bd","Type":"ContainerStarted","Data":"fc89a962a4abbaebb0621f8d21bf7b36230ad2eb0ed25b533a8c67e1accd8d6b"} Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.190976 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.192193 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bt47r" event={"ID":"3752b62f-49dd-46a7-ae34-9b6b20343a09","Type":"ContainerStarted","Data":"558c7ae1a2d2494d0a5752578c732c9e6a86618a8b63e5078ba223bedb9ce09e"} Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.194501 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-z9lkn" event={"ID":"8369f72b-14fd-40f4-864c-c8bddd0f43e8","Type":"ContainerDied","Data":"fb168c599dae50c50502a8d04510e9177e261e111ec9100752b12a727393c6bb"} Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.194633 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-z9lkn" Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.227042 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" podStartSLOduration=3.174721105 podStartE2EDuration="15.227025599s" podCreationTimestamp="2026-03-17 09:29:14 +0000 UTC" firstStartedPulling="2026-03-17 09:29:15.394284809 +0000 UTC m=+1177.495088308" lastFinishedPulling="2026-03-17 09:29:27.446589303 +0000 UTC m=+1189.547392802" observedRunningTime="2026-03-17 09:29:29.20893192 +0000 UTC m=+1191.309735419" watchObservedRunningTime="2026-03-17 09:29:29.227025599 +0000 UTC m=+1191.327829098" Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.252044 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z9lkn"] Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.253582 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-z9lkn"] Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.278474 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-988sz"] Mar 17 09:29:29 crc kubenswrapper[4813]: I0317 09:29:29.283981 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-988sz"] Mar 17 09:29:30 crc kubenswrapper[4813]: I0317 09:29:30.741671 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="63458ae3-263a-4d68-95bd-4824acb4cebd" path="/var/lib/kubelet/pods/63458ae3-263a-4d68-95bd-4824acb4cebd/volumes" Mar 17 09:29:30 crc kubenswrapper[4813]: I0317 09:29:30.742656 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8369f72b-14fd-40f4-864c-c8bddd0f43e8" path="/var/lib/kubelet/pods/8369f72b-14fd-40f4-864c-c8bddd0f43e8/volumes" Mar 17 09:29:34 crc kubenswrapper[4813]: I0317 09:29:34.898923 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" Mar 17 09:29:35 crc kubenswrapper[4813]: I0317 09:29:35.196263 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" Mar 17 09:29:35 crc kubenswrapper[4813]: I0317 09:29:35.251491 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-6cjw6"] Mar 17 09:29:35 crc kubenswrapper[4813]: I0317 09:29:35.263871 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" podUID="f0980bc5-5ad8-4479-ab3a-ea9c700428bd" containerName="dnsmasq-dns" containerID="cri-o://fc89a962a4abbaebb0621f8d21bf7b36230ad2eb0ed25b533a8c67e1accd8d6b" gracePeriod=10 Mar 17 09:29:36 crc kubenswrapper[4813]: I0317 09:29:36.272691 4813 generic.go:334] "Generic (PLEG): container finished" podID="f0980bc5-5ad8-4479-ab3a-ea9c700428bd" containerID="fc89a962a4abbaebb0621f8d21bf7b36230ad2eb0ed25b533a8c67e1accd8d6b" exitCode=0 Mar 17 09:29:36 crc kubenswrapper[4813]: I0317 09:29:36.272750 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" event={"ID":"f0980bc5-5ad8-4479-ab3a-ea9c700428bd","Type":"ContainerDied","Data":"fc89a962a4abbaebb0621f8d21bf7b36230ad2eb0ed25b533a8c67e1accd8d6b"} Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.413644 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-btlxg"] Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.416203 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-btlxg" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.418972 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.443127 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-btlxg"] Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.537844 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-6nmhn"] Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.539402 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-6nmhn" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.544368 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.549057 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-6nmhn"] Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.556576 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cbfe69c-a2c5-40bf-a8d2-e78755f060f2-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-btlxg\" (UID: \"8cbfe69c-a2c5-40bf-a8d2-e78755f060f2\") " pod="openstack/ovn-controller-metrics-btlxg" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.557104 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8cbfe69c-a2c5-40bf-a8d2-e78755f060f2-ovs-rundir\") pod \"ovn-controller-metrics-btlxg\" (UID: \"8cbfe69c-a2c5-40bf-a8d2-e78755f060f2\") " pod="openstack/ovn-controller-metrics-btlxg" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.557489 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cbfe69c-a2c5-40bf-a8d2-e78755f060f2-config\") pod \"ovn-controller-metrics-btlxg\" (UID: \"8cbfe69c-a2c5-40bf-a8d2-e78755f060f2\") " pod="openstack/ovn-controller-metrics-btlxg" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.557627 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cbfe69c-a2c5-40bf-a8d2-e78755f060f2-combined-ca-bundle\") pod \"ovn-controller-metrics-btlxg\" (UID: \"8cbfe69c-a2c5-40bf-a8d2-e78755f060f2\") " pod="openstack/ovn-controller-metrics-btlxg" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.557724 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8cbfe69c-a2c5-40bf-a8d2-e78755f060f2-ovn-rundir\") pod \"ovn-controller-metrics-btlxg\" (UID: \"8cbfe69c-a2c5-40bf-a8d2-e78755f060f2\") " pod="openstack/ovn-controller-metrics-btlxg" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.557897 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fhd2v\" (UniqueName: \"kubernetes.io/projected/8cbfe69c-a2c5-40bf-a8d2-e78755f060f2-kube-api-access-fhd2v\") pod \"ovn-controller-metrics-btlxg\" (UID: \"8cbfe69c-a2c5-40bf-a8d2-e78755f060f2\") " pod="openstack/ovn-controller-metrics-btlxg" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.659392 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cbfe69c-a2c5-40bf-a8d2-e78755f060f2-config\") pod \"ovn-controller-metrics-btlxg\" (UID: \"8cbfe69c-a2c5-40bf-a8d2-e78755f060f2\") " pod="openstack/ovn-controller-metrics-btlxg" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.659435 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cbfe69c-a2c5-40bf-a8d2-e78755f060f2-combined-ca-bundle\") pod \"ovn-controller-metrics-btlxg\" (UID: \"8cbfe69c-a2c5-40bf-a8d2-e78755f060f2\") " pod="openstack/ovn-controller-metrics-btlxg" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.659459 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mllwp\" (UniqueName: \"kubernetes.io/projected/3334a2fa-1672-4d0f-9deb-53a3036d33d8-kube-api-access-mllwp\") pod \"dnsmasq-dns-7f896c8c65-6nmhn\" (UID: \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\") " pod="openstack/dnsmasq-dns-7f896c8c65-6nmhn" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.659487 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8cbfe69c-a2c5-40bf-a8d2-e78755f060f2-ovn-rundir\") pod \"ovn-controller-metrics-btlxg\" (UID: \"8cbfe69c-a2c5-40bf-a8d2-e78755f060f2\") " pod="openstack/ovn-controller-metrics-btlxg" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.659530 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fhd2v\" (UniqueName: \"kubernetes.io/projected/8cbfe69c-a2c5-40bf-a8d2-e78755f060f2-kube-api-access-fhd2v\") pod \"ovn-controller-metrics-btlxg\" (UID: \"8cbfe69c-a2c5-40bf-a8d2-e78755f060f2\") " pod="openstack/ovn-controller-metrics-btlxg" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.659549 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3334a2fa-1672-4d0f-9deb-53a3036d33d8-config\") pod \"dnsmasq-dns-7f896c8c65-6nmhn\" (UID: \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\") " pod="openstack/dnsmasq-dns-7f896c8c65-6nmhn" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.659581 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cbfe69c-a2c5-40bf-a8d2-e78755f060f2-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-btlxg\" (UID: \"8cbfe69c-a2c5-40bf-a8d2-e78755f060f2\") " pod="openstack/ovn-controller-metrics-btlxg" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.659610 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3334a2fa-1672-4d0f-9deb-53a3036d33d8-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-6nmhn\" (UID: \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\") " pod="openstack/dnsmasq-dns-7f896c8c65-6nmhn" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.659642 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3334a2fa-1672-4d0f-9deb-53a3036d33d8-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-6nmhn\" (UID: \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\") " pod="openstack/dnsmasq-dns-7f896c8c65-6nmhn" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.659669 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8cbfe69c-a2c5-40bf-a8d2-e78755f060f2-ovs-rundir\") pod \"ovn-controller-metrics-btlxg\" (UID: \"8cbfe69c-a2c5-40bf-a8d2-e78755f060f2\") " pod="openstack/ovn-controller-metrics-btlxg" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.659934 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/8cbfe69c-a2c5-40bf-a8d2-e78755f060f2-ovs-rundir\") pod \"ovn-controller-metrics-btlxg\" (UID: \"8cbfe69c-a2c5-40bf-a8d2-e78755f060f2\") " pod="openstack/ovn-controller-metrics-btlxg" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.660195 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cbfe69c-a2c5-40bf-a8d2-e78755f060f2-config\") pod \"ovn-controller-metrics-btlxg\" (UID: \"8cbfe69c-a2c5-40bf-a8d2-e78755f060f2\") " pod="openstack/ovn-controller-metrics-btlxg" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.660288 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/8cbfe69c-a2c5-40bf-a8d2-e78755f060f2-ovn-rundir\") pod \"ovn-controller-metrics-btlxg\" (UID: \"8cbfe69c-a2c5-40bf-a8d2-e78755f060f2\") " pod="openstack/ovn-controller-metrics-btlxg" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.670004 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/8cbfe69c-a2c5-40bf-a8d2-e78755f060f2-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-btlxg\" (UID: \"8cbfe69c-a2c5-40bf-a8d2-e78755f060f2\") " pod="openstack/ovn-controller-metrics-btlxg" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.670123 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cbfe69c-a2c5-40bf-a8d2-e78755f060f2-combined-ca-bundle\") pod \"ovn-controller-metrics-btlxg\" (UID: \"8cbfe69c-a2c5-40bf-a8d2-e78755f060f2\") " pod="openstack/ovn-controller-metrics-btlxg" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.675480 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fhd2v\" (UniqueName: \"kubernetes.io/projected/8cbfe69c-a2c5-40bf-a8d2-e78755f060f2-kube-api-access-fhd2v\") pod \"ovn-controller-metrics-btlxg\" (UID: \"8cbfe69c-a2c5-40bf-a8d2-e78755f060f2\") " pod="openstack/ovn-controller-metrics-btlxg" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.741162 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-btlxg" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.761560 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3334a2fa-1672-4d0f-9deb-53a3036d33d8-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-6nmhn\" (UID: \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\") " pod="openstack/dnsmasq-dns-7f896c8c65-6nmhn" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.761721 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mllwp\" (UniqueName: \"kubernetes.io/projected/3334a2fa-1672-4d0f-9deb-53a3036d33d8-kube-api-access-mllwp\") pod \"dnsmasq-dns-7f896c8c65-6nmhn\" (UID: \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\") " pod="openstack/dnsmasq-dns-7f896c8c65-6nmhn" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.761783 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3334a2fa-1672-4d0f-9deb-53a3036d33d8-config\") pod \"dnsmasq-dns-7f896c8c65-6nmhn\" (UID: \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\") " pod="openstack/dnsmasq-dns-7f896c8c65-6nmhn" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.761832 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3334a2fa-1672-4d0f-9deb-53a3036d33d8-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-6nmhn\" (UID: \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\") " pod="openstack/dnsmasq-dns-7f896c8c65-6nmhn" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.762996 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3334a2fa-1672-4d0f-9deb-53a3036d33d8-ovsdbserver-sb\") pod \"dnsmasq-dns-7f896c8c65-6nmhn\" (UID: \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\") " pod="openstack/dnsmasq-dns-7f896c8c65-6nmhn" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.763749 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3334a2fa-1672-4d0f-9deb-53a3036d33d8-dns-svc\") pod \"dnsmasq-dns-7f896c8c65-6nmhn\" (UID: \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\") " pod="openstack/dnsmasq-dns-7f896c8c65-6nmhn" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.763844 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3334a2fa-1672-4d0f-9deb-53a3036d33d8-config\") pod \"dnsmasq-dns-7f896c8c65-6nmhn\" (UID: \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\") " pod="openstack/dnsmasq-dns-7f896c8c65-6nmhn" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.785218 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mllwp\" (UniqueName: \"kubernetes.io/projected/3334a2fa-1672-4d0f-9deb-53a3036d33d8-kube-api-access-mllwp\") pod \"dnsmasq-dns-7f896c8c65-6nmhn\" (UID: \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\") " pod="openstack/dnsmasq-dns-7f896c8c65-6nmhn" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.807024 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-6nmhn"] Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.807535 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-6nmhn" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.839099 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-rn42g"] Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.840305 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.844029 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.852554 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-rn42g"] Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.964900 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-config\") pod \"dnsmasq-dns-86db49b7ff-rn42g\" (UID: \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.964952 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-rn42g\" (UID: \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.964995 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pn7nr\" (UniqueName: \"kubernetes.io/projected/7f68b346-4221-4e2f-bb64-c43040aeb3cd-kube-api-access-pn7nr\") pod \"dnsmasq-dns-86db49b7ff-rn42g\" (UID: \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.965048 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-rn42g\" (UID: \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" Mar 17 09:29:37 crc kubenswrapper[4813]: I0317 09:29:37.965160 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-rn42g\" (UID: \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.066367 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-config\") pod \"dnsmasq-dns-86db49b7ff-rn42g\" (UID: \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.066429 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-rn42g\" (UID: \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.066471 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pn7nr\" (UniqueName: \"kubernetes.io/projected/7f68b346-4221-4e2f-bb64-c43040aeb3cd-kube-api-access-pn7nr\") pod \"dnsmasq-dns-86db49b7ff-rn42g\" (UID: \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.066517 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-rn42g\" (UID: \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.066717 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-rn42g\" (UID: \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.067614 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-dns-svc\") pod \"dnsmasq-dns-86db49b7ff-rn42g\" (UID: \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.067695 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-config\") pod \"dnsmasq-dns-86db49b7ff-rn42g\" (UID: \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.068206 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-ovsdbserver-sb\") pod \"dnsmasq-dns-86db49b7ff-rn42g\" (UID: \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.068467 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-ovsdbserver-nb\") pod \"dnsmasq-dns-86db49b7ff-rn42g\" (UID: \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.083274 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pn7nr\" (UniqueName: \"kubernetes.io/projected/7f68b346-4221-4e2f-bb64-c43040aeb3cd-kube-api-access-pn7nr\") pod \"dnsmasq-dns-86db49b7ff-rn42g\" (UID: \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\") " pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.087277 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.169443 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.169467 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0980bc5-5ad8-4479-ab3a-ea9c700428bd-config\") pod \"f0980bc5-5ad8-4479-ab3a-ea9c700428bd\" (UID: \"f0980bc5-5ad8-4479-ab3a-ea9c700428bd\") " Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.169637 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f0980bc5-5ad8-4479-ab3a-ea9c700428bd-dns-svc\") pod \"f0980bc5-5ad8-4479-ab3a-ea9c700428bd\" (UID: \"f0980bc5-5ad8-4479-ab3a-ea9c700428bd\") " Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.169691 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqnl4\" (UniqueName: \"kubernetes.io/projected/f0980bc5-5ad8-4479-ab3a-ea9c700428bd-kube-api-access-dqnl4\") pod \"f0980bc5-5ad8-4479-ab3a-ea9c700428bd\" (UID: \"f0980bc5-5ad8-4479-ab3a-ea9c700428bd\") " Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.173827 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0980bc5-5ad8-4479-ab3a-ea9c700428bd-kube-api-access-dqnl4" (OuterVolumeSpecName: "kube-api-access-dqnl4") pod "f0980bc5-5ad8-4479-ab3a-ea9c700428bd" (UID: "f0980bc5-5ad8-4479-ab3a-ea9c700428bd"). InnerVolumeSpecName "kube-api-access-dqnl4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.209202 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0980bc5-5ad8-4479-ab3a-ea9c700428bd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f0980bc5-5ad8-4479-ab3a-ea9c700428bd" (UID: "f0980bc5-5ad8-4479-ab3a-ea9c700428bd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.231018 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0980bc5-5ad8-4479-ab3a-ea9c700428bd-config" (OuterVolumeSpecName: "config") pod "f0980bc5-5ad8-4479-ab3a-ea9c700428bd" (UID: "f0980bc5-5ad8-4479-ab3a-ea9c700428bd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.271802 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f0980bc5-5ad8-4479-ab3a-ea9c700428bd-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.271844 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f0980bc5-5ad8-4479-ab3a-ea9c700428bd-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.271857 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqnl4\" (UniqueName: \"kubernetes.io/projected/f0980bc5-5ad8-4479-ab3a-ea9c700428bd-kube-api-access-dqnl4\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.287558 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" event={"ID":"f0980bc5-5ad8-4479-ab3a-ea9c700428bd","Type":"ContainerDied","Data":"0b89a1845b8f7ce226b13811970de90ee876c1ec9864ed40cdfa798aba66db91"} Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.287623 4813 scope.go:117] "RemoveContainer" containerID="fc89a962a4abbaebb0621f8d21bf7b36230ad2eb0ed25b533a8c67e1accd8d6b" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.287666 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-6cjw6" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.336458 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-6cjw6"] Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.341856 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-6cjw6"] Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.485268 4813 scope.go:117] "RemoveContainer" containerID="0c490e030ba62559a80efe6fad223b7e79fd91130b84beb558b2252202881d57" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.744617 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0980bc5-5ad8-4479-ab3a-ea9c700428bd" path="/var/lib/kubelet/pods/f0980bc5-5ad8-4479-ab3a-ea9c700428bd/volumes" Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.899852 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-6nmhn"] Mar 17 09:29:38 crc kubenswrapper[4813]: I0317 09:29:38.947121 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-btlxg"] Mar 17 09:29:39 crc kubenswrapper[4813]: W0317 09:29:39.047785 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3334a2fa_1672_4d0f_9deb_53a3036d33d8.slice/crio-d613cf2dc487d7cd63fb98a660621dac0375a752569a9aa3ff3c33b625cc2d40 WatchSource:0}: Error finding container d613cf2dc487d7cd63fb98a660621dac0375a752569a9aa3ff3c33b625cc2d40: Status 404 returned error can't find the container with id d613cf2dc487d7cd63fb98a660621dac0375a752569a9aa3ff3c33b625cc2d40 Mar 17 09:29:39 crc kubenswrapper[4813]: W0317 09:29:39.052329 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8cbfe69c_a2c5_40bf_a8d2_e78755f060f2.slice/crio-bdcbe8251e6c65b6c2aca32346e83bcbea08fb210ad909557d7c87a62c5dd57c WatchSource:0}: Error finding container bdcbe8251e6c65b6c2aca32346e83bcbea08fb210ad909557d7c87a62c5dd57c: Status 404 returned error can't find the container with id bdcbe8251e6c65b6c2aca32346e83bcbea08fb210ad909557d7c87a62c5dd57c Mar 17 09:29:39 crc kubenswrapper[4813]: I0317 09:29:39.298124 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-6nmhn" event={"ID":"3334a2fa-1672-4d0f-9deb-53a3036d33d8","Type":"ContainerStarted","Data":"d613cf2dc487d7cd63fb98a660621dac0375a752569a9aa3ff3c33b625cc2d40"} Mar 17 09:29:39 crc kubenswrapper[4813]: I0317 09:29:39.302855 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-btlxg" event={"ID":"8cbfe69c-a2c5-40bf-a8d2-e78755f060f2","Type":"ContainerStarted","Data":"bdcbe8251e6c65b6c2aca32346e83bcbea08fb210ad909557d7c87a62c5dd57c"} Mar 17 09:29:39 crc kubenswrapper[4813]: I0317 09:29:39.511789 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-rn42g"] Mar 17 09:29:39 crc kubenswrapper[4813]: W0317 09:29:39.517533 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f68b346_4221_4e2f_bb64_c43040aeb3cd.slice/crio-c230733c0c33ba3d2325c3ce1a27b75dd746d99f37079ff99315d3184e2a7e08 WatchSource:0}: Error finding container c230733c0c33ba3d2325c3ce1a27b75dd746d99f37079ff99315d3184e2a7e08: Status 404 returned error can't find the container with id c230733c0c33ba3d2325c3ce1a27b75dd746d99f37079ff99315d3184e2a7e08 Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.313996 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"59209cd3-5db8-4ddc-806e-1db457164c1f","Type":"ContainerStarted","Data":"64f16397599de8972cf374613af13195a1688ec6b89a4186a51966f796c951a4"} Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.314542 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.316941 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7e4c8a33-829e-4aec-be7c-0dad92ce3916","Type":"ContainerStarted","Data":"3c2e781de5b2b8b14705ea3592fc3dc58fd32f03294904430e734eface972599"} Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.318939 4813 generic.go:334] "Generic (PLEG): container finished" podID="3334a2fa-1672-4d0f-9deb-53a3036d33d8" containerID="5e61d0248b01ba4a5be34a7fa1da68ad7244a8880941de31b5a4baa9fe1c3a1c" exitCode=0 Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.319014 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-6nmhn" event={"ID":"3334a2fa-1672-4d0f-9deb-53a3036d33d8","Type":"ContainerDied","Data":"5e61d0248b01ba4a5be34a7fa1da68ad7244a8880941de31b5a4baa9fe1c3a1c"} Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.321175 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64","Type":"ContainerStarted","Data":"621814e2bc4a4d13b05d37567d045be23d4130507f15c3c9f85dbdb10a121b3a"} Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.324037 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vdc6c" event={"ID":"4090993b-51ce-4ce3-a6d6-a1501ab3ba05","Type":"ContainerStarted","Data":"0f46ac6327659af5688d2fe3f874e1db19cdcce9039d7a4d930607d8e924266b"} Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.324162 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-vdc6c" Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.333187 4813 generic.go:334] "Generic (PLEG): container finished" podID="3752b62f-49dd-46a7-ae34-9b6b20343a09" containerID="dcc2d25e2bee7dc566f1fa64a9fa8877f5c2ae6f76a5ea9c0f7206c1d3475288" exitCode=0 Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.333270 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bt47r" event={"ID":"3752b62f-49dd-46a7-ae34-9b6b20343a09","Type":"ContainerDied","Data":"dcc2d25e2bee7dc566f1fa64a9fa8877f5c2ae6f76a5ea9c0f7206c1d3475288"} Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.333704 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=9.434654159 podStartE2EDuration="20.333692892s" podCreationTimestamp="2026-03-17 09:29:20 +0000 UTC" firstStartedPulling="2026-03-17 09:29:28.31737417 +0000 UTC m=+1190.418177669" lastFinishedPulling="2026-03-17 09:29:39.216412903 +0000 UTC m=+1201.317216402" observedRunningTime="2026-03-17 09:29:40.32724045 +0000 UTC m=+1202.428043959" watchObservedRunningTime="2026-03-17 09:29:40.333692892 +0000 UTC m=+1202.434496391" Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.335111 4813 generic.go:334] "Generic (PLEG): container finished" podID="7f68b346-4221-4e2f-bb64-c43040aeb3cd" containerID="0511cc02d9d271cdc5e9ac7bab910f26bbf0503afb30b260655b65979ce18f31" exitCode=0 Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.335161 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" event={"ID":"7f68b346-4221-4e2f-bb64-c43040aeb3cd","Type":"ContainerDied","Data":"0511cc02d9d271cdc5e9ac7bab910f26bbf0503afb30b260655b65979ce18f31"} Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.335176 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" event={"ID":"7f68b346-4221-4e2f-bb64-c43040aeb3cd","Type":"ContainerStarted","Data":"c230733c0c33ba3d2325c3ce1a27b75dd746d99f37079ff99315d3184e2a7e08"} Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.337088 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"c43912b9-949e-4d64-ae70-e59594cc329a","Type":"ContainerStarted","Data":"eed13a577880cba2866dfd2a446f26523c0dc50cfe023249686fb8db3c9e469c"} Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.337663 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.343351 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"34e286fb-7dd1-4e85-89e3-e926b232f5a6","Type":"ContainerStarted","Data":"9847e8a7cfe7d2ef203e10291b54170073c0a66670f5d7740dbc9d7bfbcc3d83"} Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.353662 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1","Type":"ContainerStarted","Data":"c388752317317688255ada43d2dcab9c0fed03752223575a786f4b3cde7480c6"} Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.448260 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-vdc6c" podStartSLOduration=6.9707964019999995 podStartE2EDuration="16.448221064s" podCreationTimestamp="2026-03-17 09:29:24 +0000 UTC" firstStartedPulling="2026-03-17 09:29:28.330823933 +0000 UTC m=+1190.431627432" lastFinishedPulling="2026-03-17 09:29:37.808248585 +0000 UTC m=+1199.909052094" observedRunningTime="2026-03-17 09:29:40.405361307 +0000 UTC m=+1202.506164806" watchObservedRunningTime="2026-03-17 09:29:40.448221064 +0000 UTC m=+1202.549024563" Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.740847 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-6nmhn" Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.763865 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=13.343274427 podStartE2EDuration="22.763850361s" podCreationTimestamp="2026-03-17 09:29:18 +0000 UTC" firstStartedPulling="2026-03-17 09:29:28.344654958 +0000 UTC m=+1190.445458447" lastFinishedPulling="2026-03-17 09:29:37.765230882 +0000 UTC m=+1199.866034381" observedRunningTime="2026-03-17 09:29:40.534639953 +0000 UTC m=+1202.635443452" watchObservedRunningTime="2026-03-17 09:29:40.763850361 +0000 UTC m=+1202.864653860" Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.922273 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mllwp\" (UniqueName: \"kubernetes.io/projected/3334a2fa-1672-4d0f-9deb-53a3036d33d8-kube-api-access-mllwp\") pod \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\" (UID: \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\") " Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.922528 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3334a2fa-1672-4d0f-9deb-53a3036d33d8-dns-svc\") pod \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\" (UID: \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\") " Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.922712 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3334a2fa-1672-4d0f-9deb-53a3036d33d8-ovsdbserver-sb\") pod \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\" (UID: \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\") " Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.922747 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3334a2fa-1672-4d0f-9deb-53a3036d33d8-config\") pod \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\" (UID: \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\") " Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.928711 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3334a2fa-1672-4d0f-9deb-53a3036d33d8-kube-api-access-mllwp" (OuterVolumeSpecName: "kube-api-access-mllwp") pod "3334a2fa-1672-4d0f-9deb-53a3036d33d8" (UID: "3334a2fa-1672-4d0f-9deb-53a3036d33d8"). InnerVolumeSpecName "kube-api-access-mllwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.943737 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3334a2fa-1672-4d0f-9deb-53a3036d33d8-config" (OuterVolumeSpecName: "config") pod "3334a2fa-1672-4d0f-9deb-53a3036d33d8" (UID: "3334a2fa-1672-4d0f-9deb-53a3036d33d8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:29:40 crc kubenswrapper[4813]: E0317 09:29:40.952976 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/3334a2fa-1672-4d0f-9deb-53a3036d33d8-ovsdbserver-sb podName:3334a2fa-1672-4d0f-9deb-53a3036d33d8 nodeName:}" failed. No retries permitted until 2026-03-17 09:29:41.452926978 +0000 UTC m=+1203.553730477 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "ovsdbserver-sb" (UniqueName: "kubernetes.io/configmap/3334a2fa-1672-4d0f-9deb-53a3036d33d8-ovsdbserver-sb") pod "3334a2fa-1672-4d0f-9deb-53a3036d33d8" (UID: "3334a2fa-1672-4d0f-9deb-53a3036d33d8") : error deleting /var/lib/kubelet/pods/3334a2fa-1672-4d0f-9deb-53a3036d33d8/volume-subpaths: remove /var/lib/kubelet/pods/3334a2fa-1672-4d0f-9deb-53a3036d33d8/volume-subpaths: no such file or directory Mar 17 09:29:40 crc kubenswrapper[4813]: I0317 09:29:40.953204 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3334a2fa-1672-4d0f-9deb-53a3036d33d8-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3334a2fa-1672-4d0f-9deb-53a3036d33d8" (UID: "3334a2fa-1672-4d0f-9deb-53a3036d33d8"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:29:41 crc kubenswrapper[4813]: I0317 09:29:41.024375 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3334a2fa-1672-4d0f-9deb-53a3036d33d8-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:41 crc kubenswrapper[4813]: I0317 09:29:41.024406 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3334a2fa-1672-4d0f-9deb-53a3036d33d8-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:41 crc kubenswrapper[4813]: I0317 09:29:41.024420 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mllwp\" (UniqueName: \"kubernetes.io/projected/3334a2fa-1672-4d0f-9deb-53a3036d33d8-kube-api-access-mllwp\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:41 crc kubenswrapper[4813]: I0317 09:29:41.362591 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f896c8c65-6nmhn" event={"ID":"3334a2fa-1672-4d0f-9deb-53a3036d33d8","Type":"ContainerDied","Data":"d613cf2dc487d7cd63fb98a660621dac0375a752569a9aa3ff3c33b625cc2d40"} Mar 17 09:29:41 crc kubenswrapper[4813]: I0317 09:29:41.362652 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f896c8c65-6nmhn" Mar 17 09:29:41 crc kubenswrapper[4813]: I0317 09:29:41.362671 4813 scope.go:117] "RemoveContainer" containerID="5e61d0248b01ba4a5be34a7fa1da68ad7244a8880941de31b5a4baa9fe1c3a1c" Mar 17 09:29:41 crc kubenswrapper[4813]: I0317 09:29:41.365181 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bt47r" event={"ID":"3752b62f-49dd-46a7-ae34-9b6b20343a09","Type":"ContainerStarted","Data":"7e010be778d7855d4c53062a8b21f5c5a7aa09a22f264c6e06ed54b71a3fa1ec"} Mar 17 09:29:41 crc kubenswrapper[4813]: I0317 09:29:41.367115 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" event={"ID":"7f68b346-4221-4e2f-bb64-c43040aeb3cd","Type":"ContainerStarted","Data":"28d1246593453fd59a62975aeeefcc594ae26fe13729dc3b45ce495a59cb60bc"} Mar 17 09:29:41 crc kubenswrapper[4813]: I0317 09:29:41.367232 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" Mar 17 09:29:41 crc kubenswrapper[4813]: I0317 09:29:41.369144 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753","Type":"ContainerStarted","Data":"92494a89ef4e16780b97fced28f929444d04138f37ecab69d9f1fcef95b6682f"} Mar 17 09:29:41 crc kubenswrapper[4813]: I0317 09:29:41.371709 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2517b76a-309a-4c9e-88e7-6061d528acc8","Type":"ContainerStarted","Data":"d5713bfe8ab956c6c4a2dacaf5fc3b1e6cb4739a7e1e69e2bbd95c16f91dc7df"} Mar 17 09:29:41 crc kubenswrapper[4813]: I0317 09:29:41.387028 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" podStartSLOduration=4.38701172 podStartE2EDuration="4.38701172s" podCreationTimestamp="2026-03-17 09:29:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:29:41.383146739 +0000 UTC m=+1203.483950238" watchObservedRunningTime="2026-03-17 09:29:41.38701172 +0000 UTC m=+1203.487815219" Mar 17 09:29:41 crc kubenswrapper[4813]: I0317 09:29:41.531217 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3334a2fa-1672-4d0f-9deb-53a3036d33d8-ovsdbserver-sb\") pod \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\" (UID: \"3334a2fa-1672-4d0f-9deb-53a3036d33d8\") " Mar 17 09:29:41 crc kubenswrapper[4813]: I0317 09:29:41.531784 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3334a2fa-1672-4d0f-9deb-53a3036d33d8-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3334a2fa-1672-4d0f-9deb-53a3036d33d8" (UID: "3334a2fa-1672-4d0f-9deb-53a3036d33d8"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:29:41 crc kubenswrapper[4813]: I0317 09:29:41.533647 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3334a2fa-1672-4d0f-9deb-53a3036d33d8-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:41 crc kubenswrapper[4813]: I0317 09:29:41.724729 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-6nmhn"] Mar 17 09:29:41 crc kubenswrapper[4813]: I0317 09:29:41.731160 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f896c8c65-6nmhn"] Mar 17 09:29:42 crc kubenswrapper[4813]: I0317 09:29:42.739898 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3334a2fa-1672-4d0f-9deb-53a3036d33d8" path="/var/lib/kubelet/pods/3334a2fa-1672-4d0f-9deb-53a3036d33d8/volumes" Mar 17 09:29:43 crc kubenswrapper[4813]: I0317 09:29:43.390622 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"34e286fb-7dd1-4e85-89e3-e926b232f5a6","Type":"ContainerStarted","Data":"b324c1db0062539fae6e1686563fa4608d37474cfb8cfdd4665e81f62e5b05cf"} Mar 17 09:29:43 crc kubenswrapper[4813]: I0317 09:29:43.394035 4813 generic.go:334] "Generic (PLEG): container finished" podID="9ad18f78-70c7-4b1d-bda7-ed338fdcdd64" containerID="621814e2bc4a4d13b05d37567d045be23d4130507f15c3c9f85dbdb10a121b3a" exitCode=0 Mar 17 09:29:43 crc kubenswrapper[4813]: I0317 09:29:43.394100 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64","Type":"ContainerDied","Data":"621814e2bc4a4d13b05d37567d045be23d4130507f15c3c9f85dbdb10a121b3a"} Mar 17 09:29:43 crc kubenswrapper[4813]: I0317 09:29:43.397323 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"7e4c8a33-829e-4aec-be7c-0dad92ce3916","Type":"ContainerStarted","Data":"3ebe59434d017c172748ddfb893e85eb20e1de071ee6dca8883732dca25aa655"} Mar 17 09:29:43 crc kubenswrapper[4813]: I0317 09:29:43.399809 4813 generic.go:334] "Generic (PLEG): container finished" podID="95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1" containerID="c388752317317688255ada43d2dcab9c0fed03752223575a786f4b3cde7480c6" exitCode=0 Mar 17 09:29:43 crc kubenswrapper[4813]: I0317 09:29:43.399877 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1","Type":"ContainerDied","Data":"c388752317317688255ada43d2dcab9c0fed03752223575a786f4b3cde7480c6"} Mar 17 09:29:43 crc kubenswrapper[4813]: I0317 09:29:43.406495 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-bt47r" event={"ID":"3752b62f-49dd-46a7-ae34-9b6b20343a09","Type":"ContainerStarted","Data":"60cee3ec3fc3a1d812247235f32be461797b05f9df8281fa35f82522a8c51f89"} Mar 17 09:29:43 crc kubenswrapper[4813]: I0317 09:29:43.406711 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:43 crc kubenswrapper[4813]: I0317 09:29:43.406769 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:29:43 crc kubenswrapper[4813]: I0317 09:29:43.408630 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-btlxg" event={"ID":"8cbfe69c-a2c5-40bf-a8d2-e78755f060f2","Type":"ContainerStarted","Data":"744d57a1e1933cc5091ebc1d56a32826d46ce22f2cc8dffb2197e29aada4608f"} Mar 17 09:29:43 crc kubenswrapper[4813]: I0317 09:29:43.415789 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:43 crc kubenswrapper[4813]: I0317 09:29:43.415827 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:43 crc kubenswrapper[4813]: I0317 09:29:43.420433 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=6.102375438 podStartE2EDuration="20.420414912s" podCreationTimestamp="2026-03-17 09:29:23 +0000 UTC" firstStartedPulling="2026-03-17 09:29:28.539178325 +0000 UTC m=+1190.639981824" lastFinishedPulling="2026-03-17 09:29:42.857217789 +0000 UTC m=+1204.958021298" observedRunningTime="2026-03-17 09:29:43.409494489 +0000 UTC m=+1205.510297988" watchObservedRunningTime="2026-03-17 09:29:43.420414912 +0000 UTC m=+1205.521218411" Mar 17 09:29:43 crc kubenswrapper[4813]: I0317 09:29:43.440874 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-bt47r" podStartSLOduration=9.624774892 podStartE2EDuration="19.440851835s" podCreationTimestamp="2026-03-17 09:29:24 +0000 UTC" firstStartedPulling="2026-03-17 09:29:28.390580242 +0000 UTC m=+1190.491383741" lastFinishedPulling="2026-03-17 09:29:38.206657185 +0000 UTC m=+1200.307460684" observedRunningTime="2026-03-17 09:29:43.434778224 +0000 UTC m=+1205.535581733" watchObservedRunningTime="2026-03-17 09:29:43.440851835 +0000 UTC m=+1205.541655334" Mar 17 09:29:43 crc kubenswrapper[4813]: I0317 09:29:43.466486 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:43 crc kubenswrapper[4813]: I0317 09:29:43.492781 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-btlxg" podStartSLOduration=2.730615675 podStartE2EDuration="6.492760227s" podCreationTimestamp="2026-03-17 09:29:37 +0000 UTC" firstStartedPulling="2026-03-17 09:29:39.09866531 +0000 UTC m=+1201.199468809" lastFinishedPulling="2026-03-17 09:29:42.860809832 +0000 UTC m=+1204.961613361" observedRunningTime="2026-03-17 09:29:43.489836625 +0000 UTC m=+1205.590640114" watchObservedRunningTime="2026-03-17 09:29:43.492760227 +0000 UTC m=+1205.593563726" Mar 17 09:29:43 crc kubenswrapper[4813]: I0317 09:29:43.524913 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=2.710407863 podStartE2EDuration="16.524891849s" podCreationTimestamp="2026-03-17 09:29:27 +0000 UTC" firstStartedPulling="2026-03-17 09:29:29.097381662 +0000 UTC m=+1191.198185161" lastFinishedPulling="2026-03-17 09:29:42.911865648 +0000 UTC m=+1205.012669147" observedRunningTime="2026-03-17 09:29:43.516907107 +0000 UTC m=+1205.617710626" watchObservedRunningTime="2026-03-17 09:29:43.524891849 +0000 UTC m=+1205.625695348" Mar 17 09:29:44 crc kubenswrapper[4813]: I0317 09:29:44.113616 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:29:44 crc kubenswrapper[4813]: I0317 09:29:44.114675 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:29:44 crc kubenswrapper[4813]: I0317 09:29:44.216577 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Mar 17 09:29:44 crc kubenswrapper[4813]: I0317 09:29:44.416738 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"9ad18f78-70c7-4b1d-bda7-ed338fdcdd64","Type":"ContainerStarted","Data":"83761c09dd80b953c562c13865b0d9b63e362b3a55370c25eb02abaf28e12544"} Mar 17 09:29:44 crc kubenswrapper[4813]: I0317 09:29:44.418626 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1","Type":"ContainerStarted","Data":"74418a44ba73af1a526d3ae5c169edff354801db5a63e158c9184db522019b97"} Mar 17 09:29:44 crc kubenswrapper[4813]: I0317 09:29:44.465050 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=18.575624576 podStartE2EDuration="28.465029326s" podCreationTimestamp="2026-03-17 09:29:16 +0000 UTC" firstStartedPulling="2026-03-17 09:29:28.317869985 +0000 UTC m=+1190.418673484" lastFinishedPulling="2026-03-17 09:29:38.207274735 +0000 UTC m=+1200.308078234" observedRunningTime="2026-03-17 09:29:44.442411015 +0000 UTC m=+1206.543214534" watchObservedRunningTime="2026-03-17 09:29:44.465029326 +0000 UTC m=+1206.565832835" Mar 17 09:29:44 crc kubenswrapper[4813]: I0317 09:29:44.465994 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=17.448614042 podStartE2EDuration="27.465985536s" podCreationTimestamp="2026-03-17 09:29:17 +0000 UTC" firstStartedPulling="2026-03-17 09:29:28.353481766 +0000 UTC m=+1190.454285255" lastFinishedPulling="2026-03-17 09:29:38.37085325 +0000 UTC m=+1200.471656749" observedRunningTime="2026-03-17 09:29:44.460833344 +0000 UTC m=+1206.561636853" watchObservedRunningTime="2026-03-17 09:29:44.465985536 +0000 UTC m=+1206.566789055" Mar 17 09:29:44 crc kubenswrapper[4813]: I0317 09:29:44.467685 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Mar 17 09:29:44 crc kubenswrapper[4813]: E0317 09:29:44.884621 4813 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.143:32964->38.102.83.143:40999: write tcp 38.102.83.143:32964->38.102.83.143:40999: write: broken pipe Mar 17 09:29:45 crc kubenswrapper[4813]: I0317 09:29:45.204643 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.204003 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.255300 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.468116 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.659095 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Mar 17 09:29:46 crc kubenswrapper[4813]: E0317 09:29:46.659536 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3334a2fa-1672-4d0f-9deb-53a3036d33d8" containerName="init" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.659555 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3334a2fa-1672-4d0f-9deb-53a3036d33d8" containerName="init" Mar 17 09:29:46 crc kubenswrapper[4813]: E0317 09:29:46.659631 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0980bc5-5ad8-4479-ab3a-ea9c700428bd" containerName="dnsmasq-dns" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.659641 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0980bc5-5ad8-4479-ab3a-ea9c700428bd" containerName="dnsmasq-dns" Mar 17 09:29:46 crc kubenswrapper[4813]: E0317 09:29:46.659660 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0980bc5-5ad8-4479-ab3a-ea9c700428bd" containerName="init" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.659668 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0980bc5-5ad8-4479-ab3a-ea9c700428bd" containerName="init" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.659834 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="3334a2fa-1672-4d0f-9deb-53a3036d33d8" containerName="init" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.659855 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0980bc5-5ad8-4479-ab3a-ea9c700428bd" containerName="dnsmasq-dns" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.660882 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.664065 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.664251 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.664333 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.664444 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-sflj2" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.688558 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.826656 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b433a82f-1948-4fdc-a395-86a0c07fee36-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"b433a82f-1948-4fdc-a395-86a0c07fee36\") " pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.827001 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b433a82f-1948-4fdc-a395-86a0c07fee36-config\") pod \"ovn-northd-0\" (UID: \"b433a82f-1948-4fdc-a395-86a0c07fee36\") " pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.827217 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b433a82f-1948-4fdc-a395-86a0c07fee36-scripts\") pod \"ovn-northd-0\" (UID: \"b433a82f-1948-4fdc-a395-86a0c07fee36\") " pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.827309 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmmlv\" (UniqueName: \"kubernetes.io/projected/b433a82f-1948-4fdc-a395-86a0c07fee36-kube-api-access-vmmlv\") pod \"ovn-northd-0\" (UID: \"b433a82f-1948-4fdc-a395-86a0c07fee36\") " pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.827513 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/b433a82f-1948-4fdc-a395-86a0c07fee36-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"b433a82f-1948-4fdc-a395-86a0c07fee36\") " pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.827630 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b433a82f-1948-4fdc-a395-86a0c07fee36-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"b433a82f-1948-4fdc-a395-86a0c07fee36\") " pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.827742 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b433a82f-1948-4fdc-a395-86a0c07fee36-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"b433a82f-1948-4fdc-a395-86a0c07fee36\") " pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.928873 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b433a82f-1948-4fdc-a395-86a0c07fee36-scripts\") pod \"ovn-northd-0\" (UID: \"b433a82f-1948-4fdc-a395-86a0c07fee36\") " pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.928916 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmmlv\" (UniqueName: \"kubernetes.io/projected/b433a82f-1948-4fdc-a395-86a0c07fee36-kube-api-access-vmmlv\") pod \"ovn-northd-0\" (UID: \"b433a82f-1948-4fdc-a395-86a0c07fee36\") " pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.928981 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/b433a82f-1948-4fdc-a395-86a0c07fee36-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"b433a82f-1948-4fdc-a395-86a0c07fee36\") " pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.929000 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b433a82f-1948-4fdc-a395-86a0c07fee36-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"b433a82f-1948-4fdc-a395-86a0c07fee36\") " pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.929017 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b433a82f-1948-4fdc-a395-86a0c07fee36-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"b433a82f-1948-4fdc-a395-86a0c07fee36\") " pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.929047 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b433a82f-1948-4fdc-a395-86a0c07fee36-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"b433a82f-1948-4fdc-a395-86a0c07fee36\") " pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.929072 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b433a82f-1948-4fdc-a395-86a0c07fee36-config\") pod \"ovn-northd-0\" (UID: \"b433a82f-1948-4fdc-a395-86a0c07fee36\") " pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.929995 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b433a82f-1948-4fdc-a395-86a0c07fee36-config\") pod \"ovn-northd-0\" (UID: \"b433a82f-1948-4fdc-a395-86a0c07fee36\") " pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.930049 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b433a82f-1948-4fdc-a395-86a0c07fee36-scripts\") pod \"ovn-northd-0\" (UID: \"b433a82f-1948-4fdc-a395-86a0c07fee36\") " pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.930342 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/b433a82f-1948-4fdc-a395-86a0c07fee36-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"b433a82f-1948-4fdc-a395-86a0c07fee36\") " pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.935055 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/b433a82f-1948-4fdc-a395-86a0c07fee36-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"b433a82f-1948-4fdc-a395-86a0c07fee36\") " pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.935472 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/b433a82f-1948-4fdc-a395-86a0c07fee36-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"b433a82f-1948-4fdc-a395-86a0c07fee36\") " pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.936086 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b433a82f-1948-4fdc-a395-86a0c07fee36-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"b433a82f-1948-4fdc-a395-86a0c07fee36\") " pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.946882 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmmlv\" (UniqueName: \"kubernetes.io/projected/b433a82f-1948-4fdc-a395-86a0c07fee36-kube-api-access-vmmlv\") pod \"ovn-northd-0\" (UID: \"b433a82f-1948-4fdc-a395-86a0c07fee36\") " pod="openstack/ovn-northd-0" Mar 17 09:29:46 crc kubenswrapper[4813]: I0317 09:29:46.986947 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Mar 17 09:29:47 crc kubenswrapper[4813]: I0317 09:29:47.390069 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Mar 17 09:29:47 crc kubenswrapper[4813]: I0317 09:29:47.390160 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Mar 17 09:29:47 crc kubenswrapper[4813]: I0317 09:29:47.420156 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Mar 17 09:29:47 crc kubenswrapper[4813]: W0317 09:29:47.426924 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb433a82f_1948_4fdc_a395_86a0c07fee36.slice/crio-769cfc531fb1e4110c72882ebd2efd2f009d628d9e33a90a00f584155ca0ec29 WatchSource:0}: Error finding container 769cfc531fb1e4110c72882ebd2efd2f009d628d9e33a90a00f584155ca0ec29: Status 404 returned error can't find the container with id 769cfc531fb1e4110c72882ebd2efd2f009d628d9e33a90a00f584155ca0ec29 Mar 17 09:29:47 crc kubenswrapper[4813]: I0317 09:29:47.448211 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"b433a82f-1948-4fdc-a395-86a0c07fee36","Type":"ContainerStarted","Data":"769cfc531fb1e4110c72882ebd2efd2f009d628d9e33a90a00f584155ca0ec29"} Mar 17 09:29:48 crc kubenswrapper[4813]: I0317 09:29:48.171518 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" Mar 17 09:29:48 crc kubenswrapper[4813]: I0317 09:29:48.228367 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ksnd5"] Mar 17 09:29:48 crc kubenswrapper[4813]: I0317 09:29:48.228984 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" podUID="fc31e29a-535b-401c-bc91-b146d3a53617" containerName="dnsmasq-dns" containerID="cri-o://062968e76b36dc557aa11d7a139d611432719617f05f9c248fc9754be1a25125" gracePeriod=10 Mar 17 09:29:48 crc kubenswrapper[4813]: I0317 09:29:48.470586 4813 generic.go:334] "Generic (PLEG): container finished" podID="fc31e29a-535b-401c-bc91-b146d3a53617" containerID="062968e76b36dc557aa11d7a139d611432719617f05f9c248fc9754be1a25125" exitCode=0 Mar 17 09:29:48 crc kubenswrapper[4813]: I0317 09:29:48.471453 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" event={"ID":"fc31e29a-535b-401c-bc91-b146d3a53617","Type":"ContainerDied","Data":"062968e76b36dc557aa11d7a139d611432719617f05f9c248fc9754be1a25125"} Mar 17 09:29:48 crc kubenswrapper[4813]: I0317 09:29:48.884568 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:48 crc kubenswrapper[4813]: I0317 09:29:48.884639 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:48 crc kubenswrapper[4813]: I0317 09:29:48.991114 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:49 crc kubenswrapper[4813]: I0317 09:29:49.125377 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" Mar 17 09:29:49 crc kubenswrapper[4813]: I0317 09:29:49.263254 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hlc8m\" (UniqueName: \"kubernetes.io/projected/fc31e29a-535b-401c-bc91-b146d3a53617-kube-api-access-hlc8m\") pod \"fc31e29a-535b-401c-bc91-b146d3a53617\" (UID: \"fc31e29a-535b-401c-bc91-b146d3a53617\") " Mar 17 09:29:49 crc kubenswrapper[4813]: I0317 09:29:49.263294 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc31e29a-535b-401c-bc91-b146d3a53617-config\") pod \"fc31e29a-535b-401c-bc91-b146d3a53617\" (UID: \"fc31e29a-535b-401c-bc91-b146d3a53617\") " Mar 17 09:29:49 crc kubenswrapper[4813]: I0317 09:29:49.263328 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc31e29a-535b-401c-bc91-b146d3a53617-dns-svc\") pod \"fc31e29a-535b-401c-bc91-b146d3a53617\" (UID: \"fc31e29a-535b-401c-bc91-b146d3a53617\") " Mar 17 09:29:49 crc kubenswrapper[4813]: I0317 09:29:49.266062 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc31e29a-535b-401c-bc91-b146d3a53617-kube-api-access-hlc8m" (OuterVolumeSpecName: "kube-api-access-hlc8m") pod "fc31e29a-535b-401c-bc91-b146d3a53617" (UID: "fc31e29a-535b-401c-bc91-b146d3a53617"). InnerVolumeSpecName "kube-api-access-hlc8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:29:49 crc kubenswrapper[4813]: I0317 09:29:49.304774 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc31e29a-535b-401c-bc91-b146d3a53617-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fc31e29a-535b-401c-bc91-b146d3a53617" (UID: "fc31e29a-535b-401c-bc91-b146d3a53617"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:29:49 crc kubenswrapper[4813]: I0317 09:29:49.311454 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fc31e29a-535b-401c-bc91-b146d3a53617-config" (OuterVolumeSpecName: "config") pod "fc31e29a-535b-401c-bc91-b146d3a53617" (UID: "fc31e29a-535b-401c-bc91-b146d3a53617"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:29:49 crc kubenswrapper[4813]: I0317 09:29:49.365665 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hlc8m\" (UniqueName: \"kubernetes.io/projected/fc31e29a-535b-401c-bc91-b146d3a53617-kube-api-access-hlc8m\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:49 crc kubenswrapper[4813]: I0317 09:29:49.365727 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fc31e29a-535b-401c-bc91-b146d3a53617-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:49 crc kubenswrapper[4813]: I0317 09:29:49.365748 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fc31e29a-535b-401c-bc91-b146d3a53617-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:49 crc kubenswrapper[4813]: I0317 09:29:49.483708 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"b433a82f-1948-4fdc-a395-86a0c07fee36","Type":"ContainerStarted","Data":"2cf408c57e67e2e190a9099b70210f2ded72427653dda5fb4ac852fd769ceea8"} Mar 17 09:29:49 crc kubenswrapper[4813]: I0317 09:29:49.483752 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"b433a82f-1948-4fdc-a395-86a0c07fee36","Type":"ContainerStarted","Data":"d6890015d6b16f8300d02a010f31461406d24df7b98ff12a2f3d15cc2d3bf29a"} Mar 17 09:29:49 crc kubenswrapper[4813]: I0317 09:29:49.483864 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Mar 17 09:29:49 crc kubenswrapper[4813]: I0317 09:29:49.486037 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" event={"ID":"fc31e29a-535b-401c-bc91-b146d3a53617","Type":"ContainerDied","Data":"669e972f9a41e2309b46825715b397c68337c9f6a866fd690447fdfb95bb1b32"} Mar 17 09:29:49 crc kubenswrapper[4813]: I0317 09:29:49.486063 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-ksnd5" Mar 17 09:29:49 crc kubenswrapper[4813]: I0317 09:29:49.486087 4813 scope.go:117] "RemoveContainer" containerID="062968e76b36dc557aa11d7a139d611432719617f05f9c248fc9754be1a25125" Mar 17 09:29:49 crc kubenswrapper[4813]: I0317 09:29:49.507830 4813 scope.go:117] "RemoveContainer" containerID="14dbda82b6001173ff09d165849f6a30a02b46b5c74614a824fb80ca14c07ff5" Mar 17 09:29:49 crc kubenswrapper[4813]: I0317 09:29:49.522430 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=1.993780646 podStartE2EDuration="3.522408813s" podCreationTimestamp="2026-03-17 09:29:46 +0000 UTC" firstStartedPulling="2026-03-17 09:29:47.430571143 +0000 UTC m=+1209.531374672" lastFinishedPulling="2026-03-17 09:29:48.95919934 +0000 UTC m=+1211.060002839" observedRunningTime="2026-03-17 09:29:49.508257648 +0000 UTC m=+1211.609061157" watchObservedRunningTime="2026-03-17 09:29:49.522408813 +0000 UTC m=+1211.623212312" Mar 17 09:29:49 crc kubenswrapper[4813]: I0317 09:29:49.536396 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ksnd5"] Mar 17 09:29:49 crc kubenswrapper[4813]: I0317 09:29:49.544262 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-ksnd5"] Mar 17 09:29:49 crc kubenswrapper[4813]: I0317 09:29:49.586735 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Mar 17 09:29:50 crc kubenswrapper[4813]: I0317 09:29:50.163971 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Mar 17 09:29:50 crc kubenswrapper[4813]: I0317 09:29:50.276251 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Mar 17 09:29:50 crc kubenswrapper[4813]: I0317 09:29:50.560554 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6ee7-account-create-update-ffh6m"] Mar 17 09:29:50 crc kubenswrapper[4813]: E0317 09:29:50.561107 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc31e29a-535b-401c-bc91-b146d3a53617" containerName="dnsmasq-dns" Mar 17 09:29:50 crc kubenswrapper[4813]: I0317 09:29:50.561127 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc31e29a-535b-401c-bc91-b146d3a53617" containerName="dnsmasq-dns" Mar 17 09:29:50 crc kubenswrapper[4813]: E0317 09:29:50.561205 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc31e29a-535b-401c-bc91-b146d3a53617" containerName="init" Mar 17 09:29:50 crc kubenswrapper[4813]: I0317 09:29:50.561220 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc31e29a-535b-401c-bc91-b146d3a53617" containerName="init" Mar 17 09:29:50 crc kubenswrapper[4813]: I0317 09:29:50.561522 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc31e29a-535b-401c-bc91-b146d3a53617" containerName="dnsmasq-dns" Mar 17 09:29:50 crc kubenswrapper[4813]: I0317 09:29:50.562332 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6ee7-account-create-update-ffh6m" Mar 17 09:29:50 crc kubenswrapper[4813]: I0317 09:29:50.566394 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Mar 17 09:29:50 crc kubenswrapper[4813]: I0317 09:29:50.573025 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6ee7-account-create-update-ffh6m"] Mar 17 09:29:50 crc kubenswrapper[4813]: I0317 09:29:50.688831 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twx4n\" (UniqueName: \"kubernetes.io/projected/35511914-5b8f-45e4-87cb-cd174e467dcc-kube-api-access-twx4n\") pod \"placement-6ee7-account-create-update-ffh6m\" (UID: \"35511914-5b8f-45e4-87cb-cd174e467dcc\") " pod="openstack/placement-6ee7-account-create-update-ffh6m" Mar 17 09:29:50 crc kubenswrapper[4813]: I0317 09:29:50.689121 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35511914-5b8f-45e4-87cb-cd174e467dcc-operator-scripts\") pod \"placement-6ee7-account-create-update-ffh6m\" (UID: \"35511914-5b8f-45e4-87cb-cd174e467dcc\") " pod="openstack/placement-6ee7-account-create-update-ffh6m" Mar 17 09:29:50 crc kubenswrapper[4813]: I0317 09:29:50.739955 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc31e29a-535b-401c-bc91-b146d3a53617" path="/var/lib/kubelet/pods/fc31e29a-535b-401c-bc91-b146d3a53617/volumes" Mar 17 09:29:50 crc kubenswrapper[4813]: I0317 09:29:50.790340 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twx4n\" (UniqueName: \"kubernetes.io/projected/35511914-5b8f-45e4-87cb-cd174e467dcc-kube-api-access-twx4n\") pod \"placement-6ee7-account-create-update-ffh6m\" (UID: \"35511914-5b8f-45e4-87cb-cd174e467dcc\") " pod="openstack/placement-6ee7-account-create-update-ffh6m" Mar 17 09:29:50 crc kubenswrapper[4813]: I0317 09:29:50.790416 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35511914-5b8f-45e4-87cb-cd174e467dcc-operator-scripts\") pod \"placement-6ee7-account-create-update-ffh6m\" (UID: \"35511914-5b8f-45e4-87cb-cd174e467dcc\") " pod="openstack/placement-6ee7-account-create-update-ffh6m" Mar 17 09:29:50 crc kubenswrapper[4813]: I0317 09:29:50.791119 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35511914-5b8f-45e4-87cb-cd174e467dcc-operator-scripts\") pod \"placement-6ee7-account-create-update-ffh6m\" (UID: \"35511914-5b8f-45e4-87cb-cd174e467dcc\") " pod="openstack/placement-6ee7-account-create-update-ffh6m" Mar 17 09:29:50 crc kubenswrapper[4813]: I0317 09:29:50.816433 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twx4n\" (UniqueName: \"kubernetes.io/projected/35511914-5b8f-45e4-87cb-cd174e467dcc-kube-api-access-twx4n\") pod \"placement-6ee7-account-create-update-ffh6m\" (UID: \"35511914-5b8f-45e4-87cb-cd174e467dcc\") " pod="openstack/placement-6ee7-account-create-update-ffh6m" Mar 17 09:29:50 crc kubenswrapper[4813]: I0317 09:29:50.885094 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6ee7-account-create-update-ffh6m" Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.301383 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.322507 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-b25vw"] Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.323873 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-b25vw" Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.346020 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-b25vw"] Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.359766 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6ee7-account-create-update-ffh6m"] Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.400688 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-dns-svc\") pod \"dnsmasq-dns-698758b865-b25vw\" (UID: \"8552c885-2651-4424-b0d6-f75a5b7487dc\") " pod="openstack/dnsmasq-dns-698758b865-b25vw" Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.400797 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-config\") pod \"dnsmasq-dns-698758b865-b25vw\" (UID: \"8552c885-2651-4424-b0d6-f75a5b7487dc\") " pod="openstack/dnsmasq-dns-698758b865-b25vw" Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.400850 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-b25vw\" (UID: \"8552c885-2651-4424-b0d6-f75a5b7487dc\") " pod="openstack/dnsmasq-dns-698758b865-b25vw" Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.400926 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-b25vw\" (UID: \"8552c885-2651-4424-b0d6-f75a5b7487dc\") " pod="openstack/dnsmasq-dns-698758b865-b25vw" Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.400956 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2f74\" (UniqueName: \"kubernetes.io/projected/8552c885-2651-4424-b0d6-f75a5b7487dc-kube-api-access-c2f74\") pod \"dnsmasq-dns-698758b865-b25vw\" (UID: \"8552c885-2651-4424-b0d6-f75a5b7487dc\") " pod="openstack/dnsmasq-dns-698758b865-b25vw" Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.503734 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-dns-svc\") pod \"dnsmasq-dns-698758b865-b25vw\" (UID: \"8552c885-2651-4424-b0d6-f75a5b7487dc\") " pod="openstack/dnsmasq-dns-698758b865-b25vw" Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.503827 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-config\") pod \"dnsmasq-dns-698758b865-b25vw\" (UID: \"8552c885-2651-4424-b0d6-f75a5b7487dc\") " pod="openstack/dnsmasq-dns-698758b865-b25vw" Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.503865 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-b25vw\" (UID: \"8552c885-2651-4424-b0d6-f75a5b7487dc\") " pod="openstack/dnsmasq-dns-698758b865-b25vw" Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.503906 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-b25vw\" (UID: \"8552c885-2651-4424-b0d6-f75a5b7487dc\") " pod="openstack/dnsmasq-dns-698758b865-b25vw" Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.503927 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2f74\" (UniqueName: \"kubernetes.io/projected/8552c885-2651-4424-b0d6-f75a5b7487dc-kube-api-access-c2f74\") pod \"dnsmasq-dns-698758b865-b25vw\" (UID: \"8552c885-2651-4424-b0d6-f75a5b7487dc\") " pod="openstack/dnsmasq-dns-698758b865-b25vw" Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.504772 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-dns-svc\") pod \"dnsmasq-dns-698758b865-b25vw\" (UID: \"8552c885-2651-4424-b0d6-f75a5b7487dc\") " pod="openstack/dnsmasq-dns-698758b865-b25vw" Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.504917 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-config\") pod \"dnsmasq-dns-698758b865-b25vw\" (UID: \"8552c885-2651-4424-b0d6-f75a5b7487dc\") " pod="openstack/dnsmasq-dns-698758b865-b25vw" Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.505427 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-b25vw\" (UID: \"8552c885-2651-4424-b0d6-f75a5b7487dc\") " pod="openstack/dnsmasq-dns-698758b865-b25vw" Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.505473 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-b25vw\" (UID: \"8552c885-2651-4424-b0d6-f75a5b7487dc\") " pod="openstack/dnsmasq-dns-698758b865-b25vw" Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.511818 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6ee7-account-create-update-ffh6m" event={"ID":"35511914-5b8f-45e4-87cb-cd174e467dcc","Type":"ContainerStarted","Data":"6c93c91cbe1e2fa43f2d31ddf6e1f46210bae5b23a4363e6ba9073adbf759542"} Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.529991 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2f74\" (UniqueName: \"kubernetes.io/projected/8552c885-2651-4424-b0d6-f75a5b7487dc-kube-api-access-c2f74\") pod \"dnsmasq-dns-698758b865-b25vw\" (UID: \"8552c885-2651-4424-b0d6-f75a5b7487dc\") " pod="openstack/dnsmasq-dns-698758b865-b25vw" Mar 17 09:29:51 crc kubenswrapper[4813]: I0317 09:29:51.643373 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-b25vw" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.062999 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-b25vw"] Mar 17 09:29:52 crc kubenswrapper[4813]: W0317 09:29:52.069999 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8552c885_2651_4424_b0d6_f75a5b7487dc.slice/crio-dd9b580dc06a35b506c8734f8284b9ced2e59287231eac3456b7472045d76499 WatchSource:0}: Error finding container dd9b580dc06a35b506c8734f8284b9ced2e59287231eac3456b7472045d76499: Status 404 returned error can't find the container with id dd9b580dc06a35b506c8734f8284b9ced2e59287231eac3456b7472045d76499 Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.475804 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.480842 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.482934 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.483250 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.483424 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.483853 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-l26lm" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.499883 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.520339 4813 generic.go:334] "Generic (PLEG): container finished" podID="35511914-5b8f-45e4-87cb-cd174e467dcc" containerID="26fce0a124538082a35190759f3217c42a5039ec4a698e529f898ce6480af7e7" exitCode=0 Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.520388 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6ee7-account-create-update-ffh6m" event={"ID":"35511914-5b8f-45e4-87cb-cd174e467dcc","Type":"ContainerDied","Data":"26fce0a124538082a35190759f3217c42a5039ec4a698e529f898ce6480af7e7"} Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.521635 4813 generic.go:334] "Generic (PLEG): container finished" podID="8552c885-2651-4424-b0d6-f75a5b7487dc" containerID="8cdd5e139009b415620c6c446d0d0577abceb05756f4f96280bcb870570be28e" exitCode=0 Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.521684 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-b25vw" event={"ID":"8552c885-2651-4424-b0d6-f75a5b7487dc","Type":"ContainerDied","Data":"8cdd5e139009b415620c6c446d0d0577abceb05756f4f96280bcb870570be28e"} Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.521713 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-b25vw" event={"ID":"8552c885-2651-4424-b0d6-f75a5b7487dc","Type":"ContainerStarted","Data":"dd9b580dc06a35b506c8734f8284b9ced2e59287231eac3456b7472045d76499"} Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.524802 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.524871 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-etc-swift\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.524931 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-cache\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.524988 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rlh7g\" (UniqueName: \"kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-kube-api-access-rlh7g\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.525016 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-lock\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.525041 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.625910 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-etc-swift\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.626004 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-cache\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.626057 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rlh7g\" (UniqueName: \"kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-kube-api-access-rlh7g\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.626082 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-lock\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.626115 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.626180 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:29:52 crc kubenswrapper[4813]: E0317 09:29:52.626649 4813 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 17 09:29:52 crc kubenswrapper[4813]: E0317 09:29:52.626684 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 17 09:29:52 crc kubenswrapper[4813]: E0317 09:29:52.626756 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-etc-swift podName:76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51 nodeName:}" failed. No retries permitted until 2026-03-17 09:29:53.126730947 +0000 UTC m=+1215.227534546 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-etc-swift") pod "swift-storage-0" (UID: "76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51") : configmap "swift-ring-files" not found Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.626759 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/swift-storage-0" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.626942 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-cache\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.627177 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-lock\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.630573 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-combined-ca-bundle\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.642467 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rlh7g\" (UniqueName: \"kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-kube-api-access-rlh7g\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:29:52 crc kubenswrapper[4813]: I0317 09:29:52.649720 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.003579 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-wcmzb"] Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.006049 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.009994 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.010491 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.010718 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.033971 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e94939bc-6595-45af-94b8-148e850b86ff-dispersionconf\") pod \"swift-ring-rebalance-wcmzb\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.034034 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svr45\" (UniqueName: \"kubernetes.io/projected/e94939bc-6595-45af-94b8-148e850b86ff-kube-api-access-svr45\") pod \"swift-ring-rebalance-wcmzb\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.034068 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e94939bc-6595-45af-94b8-148e850b86ff-swiftconf\") pod \"swift-ring-rebalance-wcmzb\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.034090 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e94939bc-6595-45af-94b8-148e850b86ff-etc-swift\") pod \"swift-ring-rebalance-wcmzb\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.034151 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e94939bc-6595-45af-94b8-148e850b86ff-ring-data-devices\") pod \"swift-ring-rebalance-wcmzb\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.034508 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e94939bc-6595-45af-94b8-148e850b86ff-combined-ca-bundle\") pod \"swift-ring-rebalance-wcmzb\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.034652 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e94939bc-6595-45af-94b8-148e850b86ff-scripts\") pod \"swift-ring-rebalance-wcmzb\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.068431 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-rrfrc"] Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.070443 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.085951 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-wcmzb"] Mar 17 09:29:53 crc kubenswrapper[4813]: E0317 09:29:53.086820 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle dispersionconf etc-swift kube-api-access-svr45 ring-data-devices scripts swiftconf], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/swift-ring-rebalance-wcmzb" podUID="e94939bc-6595-45af-94b8-148e850b86ff" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.094377 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-rrfrc"] Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.100083 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-wcmzb"] Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.136655 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8jhf\" (UniqueName: \"kubernetes.io/projected/625a6e48-f069-49a9-b11a-342e12dffece-kube-api-access-r8jhf\") pod \"swift-ring-rebalance-rrfrc\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.136735 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/625a6e48-f069-49a9-b11a-342e12dffece-ring-data-devices\") pod \"swift-ring-rebalance-rrfrc\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.136777 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e94939bc-6595-45af-94b8-148e850b86ff-ring-data-devices\") pod \"swift-ring-rebalance-wcmzb\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.136825 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/625a6e48-f069-49a9-b11a-342e12dffece-swiftconf\") pod \"swift-ring-rebalance-rrfrc\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.136865 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-etc-swift\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.136927 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/625a6e48-f069-49a9-b11a-342e12dffece-dispersionconf\") pod \"swift-ring-rebalance-rrfrc\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.136952 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/625a6e48-f069-49a9-b11a-342e12dffece-combined-ca-bundle\") pod \"swift-ring-rebalance-rrfrc\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.136982 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e94939bc-6595-45af-94b8-148e850b86ff-combined-ca-bundle\") pod \"swift-ring-rebalance-wcmzb\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.137011 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e94939bc-6595-45af-94b8-148e850b86ff-scripts\") pod \"swift-ring-rebalance-wcmzb\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.137054 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/625a6e48-f069-49a9-b11a-342e12dffece-etc-swift\") pod \"swift-ring-rebalance-rrfrc\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.137086 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e94939bc-6595-45af-94b8-148e850b86ff-dispersionconf\") pod \"swift-ring-rebalance-wcmzb\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.137124 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svr45\" (UniqueName: \"kubernetes.io/projected/e94939bc-6595-45af-94b8-148e850b86ff-kube-api-access-svr45\") pod \"swift-ring-rebalance-wcmzb\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.137149 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/625a6e48-f069-49a9-b11a-342e12dffece-scripts\") pod \"swift-ring-rebalance-rrfrc\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.137176 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e94939bc-6595-45af-94b8-148e850b86ff-swiftconf\") pod \"swift-ring-rebalance-wcmzb\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.137198 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e94939bc-6595-45af-94b8-148e850b86ff-etc-swift\") pod \"swift-ring-rebalance-wcmzb\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.137748 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e94939bc-6595-45af-94b8-148e850b86ff-etc-swift\") pod \"swift-ring-rebalance-wcmzb\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.138348 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e94939bc-6595-45af-94b8-148e850b86ff-scripts\") pod \"swift-ring-rebalance-wcmzb\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.138522 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e94939bc-6595-45af-94b8-148e850b86ff-ring-data-devices\") pod \"swift-ring-rebalance-wcmzb\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: E0317 09:29:53.138581 4813 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 17 09:29:53 crc kubenswrapper[4813]: E0317 09:29:53.138786 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 17 09:29:53 crc kubenswrapper[4813]: E0317 09:29:53.138919 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-etc-swift podName:76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51 nodeName:}" failed. No retries permitted until 2026-03-17 09:29:54.138895585 +0000 UTC m=+1216.239699124 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-etc-swift") pod "swift-storage-0" (UID: "76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51") : configmap "swift-ring-files" not found Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.142471 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e94939bc-6595-45af-94b8-148e850b86ff-dispersionconf\") pod \"swift-ring-rebalance-wcmzb\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.150255 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e94939bc-6595-45af-94b8-148e850b86ff-combined-ca-bundle\") pod \"swift-ring-rebalance-wcmzb\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.154373 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svr45\" (UniqueName: \"kubernetes.io/projected/e94939bc-6595-45af-94b8-148e850b86ff-kube-api-access-svr45\") pod \"swift-ring-rebalance-wcmzb\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.158713 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e94939bc-6595-45af-94b8-148e850b86ff-swiftconf\") pod \"swift-ring-rebalance-wcmzb\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.238765 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/625a6e48-f069-49a9-b11a-342e12dffece-scripts\") pod \"swift-ring-rebalance-rrfrc\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.240065 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8jhf\" (UniqueName: \"kubernetes.io/projected/625a6e48-f069-49a9-b11a-342e12dffece-kube-api-access-r8jhf\") pod \"swift-ring-rebalance-rrfrc\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.240542 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/625a6e48-f069-49a9-b11a-342e12dffece-ring-data-devices\") pod \"swift-ring-rebalance-rrfrc\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.241330 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/625a6e48-f069-49a9-b11a-342e12dffece-swiftconf\") pod \"swift-ring-rebalance-rrfrc\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.241234 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/625a6e48-f069-49a9-b11a-342e12dffece-ring-data-devices\") pod \"swift-ring-rebalance-rrfrc\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.239983 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/625a6e48-f069-49a9-b11a-342e12dffece-scripts\") pod \"swift-ring-rebalance-rrfrc\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.241973 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/625a6e48-f069-49a9-b11a-342e12dffece-dispersionconf\") pod \"swift-ring-rebalance-rrfrc\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.242088 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/625a6e48-f069-49a9-b11a-342e12dffece-combined-ca-bundle\") pod \"swift-ring-rebalance-rrfrc\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.242195 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/625a6e48-f069-49a9-b11a-342e12dffece-etc-swift\") pod \"swift-ring-rebalance-rrfrc\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.242646 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/625a6e48-f069-49a9-b11a-342e12dffece-etc-swift\") pod \"swift-ring-rebalance-rrfrc\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.244907 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/625a6e48-f069-49a9-b11a-342e12dffece-swiftconf\") pod \"swift-ring-rebalance-rrfrc\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.245802 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/625a6e48-f069-49a9-b11a-342e12dffece-dispersionconf\") pod \"swift-ring-rebalance-rrfrc\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.248281 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/625a6e48-f069-49a9-b11a-342e12dffece-combined-ca-bundle\") pod \"swift-ring-rebalance-rrfrc\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.260717 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8jhf\" (UniqueName: \"kubernetes.io/projected/625a6e48-f069-49a9-b11a-342e12dffece-kube-api-access-r8jhf\") pod \"swift-ring-rebalance-rrfrc\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.388740 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.533550 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-b25vw" event={"ID":"8552c885-2651-4424-b0d6-f75a5b7487dc","Type":"ContainerStarted","Data":"0f2903c7e92e7817c2c04a7025fca74674207ad263f5ddacaf07b970622b3116"} Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.533749 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.533970 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-b25vw" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.555341 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.560767 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-b25vw" podStartSLOduration=2.560750723 podStartE2EDuration="2.560750723s" podCreationTimestamp="2026-03-17 09:29:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:29:53.559592766 +0000 UTC m=+1215.660396285" watchObservedRunningTime="2026-03-17 09:29:53.560750723 +0000 UTC m=+1215.661554222" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.648752 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e94939bc-6595-45af-94b8-148e850b86ff-ring-data-devices\") pod \"e94939bc-6595-45af-94b8-148e850b86ff\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.648790 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e94939bc-6595-45af-94b8-148e850b86ff-scripts\") pod \"e94939bc-6595-45af-94b8-148e850b86ff\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.648820 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svr45\" (UniqueName: \"kubernetes.io/projected/e94939bc-6595-45af-94b8-148e850b86ff-kube-api-access-svr45\") pod \"e94939bc-6595-45af-94b8-148e850b86ff\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.648872 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e94939bc-6595-45af-94b8-148e850b86ff-dispersionconf\") pod \"e94939bc-6595-45af-94b8-148e850b86ff\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.648897 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e94939bc-6595-45af-94b8-148e850b86ff-etc-swift\") pod \"e94939bc-6595-45af-94b8-148e850b86ff\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.648922 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e94939bc-6595-45af-94b8-148e850b86ff-swiftconf\") pod \"e94939bc-6595-45af-94b8-148e850b86ff\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.648944 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e94939bc-6595-45af-94b8-148e850b86ff-combined-ca-bundle\") pod \"e94939bc-6595-45af-94b8-148e850b86ff\" (UID: \"e94939bc-6595-45af-94b8-148e850b86ff\") " Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.649464 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e94939bc-6595-45af-94b8-148e850b86ff-scripts" (OuterVolumeSpecName: "scripts") pod "e94939bc-6595-45af-94b8-148e850b86ff" (UID: "e94939bc-6595-45af-94b8-148e850b86ff"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.649525 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e94939bc-6595-45af-94b8-148e850b86ff-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "e94939bc-6595-45af-94b8-148e850b86ff" (UID: "e94939bc-6595-45af-94b8-148e850b86ff"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.649861 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e94939bc-6595-45af-94b8-148e850b86ff-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "e94939bc-6595-45af-94b8-148e850b86ff" (UID: "e94939bc-6595-45af-94b8-148e850b86ff"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.654017 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e94939bc-6595-45af-94b8-148e850b86ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e94939bc-6595-45af-94b8-148e850b86ff" (UID: "e94939bc-6595-45af-94b8-148e850b86ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.654745 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e94939bc-6595-45af-94b8-148e850b86ff-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "e94939bc-6595-45af-94b8-148e850b86ff" (UID: "e94939bc-6595-45af-94b8-148e850b86ff"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.655152 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e94939bc-6595-45af-94b8-148e850b86ff-kube-api-access-svr45" (OuterVolumeSpecName: "kube-api-access-svr45") pod "e94939bc-6595-45af-94b8-148e850b86ff" (UID: "e94939bc-6595-45af-94b8-148e850b86ff"). InnerVolumeSpecName "kube-api-access-svr45". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.655839 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e94939bc-6595-45af-94b8-148e850b86ff-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "e94939bc-6595-45af-94b8-148e850b86ff" (UID: "e94939bc-6595-45af-94b8-148e850b86ff"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.751098 4813 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/e94939bc-6595-45af-94b8-148e850b86ff-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.751134 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e94939bc-6595-45af-94b8-148e850b86ff-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.751145 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svr45\" (UniqueName: \"kubernetes.io/projected/e94939bc-6595-45af-94b8-148e850b86ff-kube-api-access-svr45\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.751154 4813 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/e94939bc-6595-45af-94b8-148e850b86ff-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.751162 4813 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/e94939bc-6595-45af-94b8-148e850b86ff-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.751170 4813 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/e94939bc-6595-45af-94b8-148e850b86ff-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.751179 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e94939bc-6595-45af-94b8-148e850b86ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.803310 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6ee7-account-create-update-ffh6m" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.852206 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35511914-5b8f-45e4-87cb-cd174e467dcc-operator-scripts\") pod \"35511914-5b8f-45e4-87cb-cd174e467dcc\" (UID: \"35511914-5b8f-45e4-87cb-cd174e467dcc\") " Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.852573 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twx4n\" (UniqueName: \"kubernetes.io/projected/35511914-5b8f-45e4-87cb-cd174e467dcc-kube-api-access-twx4n\") pod \"35511914-5b8f-45e4-87cb-cd174e467dcc\" (UID: \"35511914-5b8f-45e4-87cb-cd174e467dcc\") " Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.852768 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/35511914-5b8f-45e4-87cb-cd174e467dcc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "35511914-5b8f-45e4-87cb-cd174e467dcc" (UID: "35511914-5b8f-45e4-87cb-cd174e467dcc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.853072 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/35511914-5b8f-45e4-87cb-cd174e467dcc-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.861310 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35511914-5b8f-45e4-87cb-cd174e467dcc-kube-api-access-twx4n" (OuterVolumeSpecName: "kube-api-access-twx4n") pod "35511914-5b8f-45e4-87cb-cd174e467dcc" (UID: "35511914-5b8f-45e4-87cb-cd174e467dcc"). InnerVolumeSpecName "kube-api-access-twx4n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.866819 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-rrfrc"] Mar 17 09:29:53 crc kubenswrapper[4813]: W0317 09:29:53.872360 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod625a6e48_f069_49a9_b11a_342e12dffece.slice/crio-f543e5c178e08b935eaf33bcc1d7ea23c030cfb716926fe5ca79788df814c77c WatchSource:0}: Error finding container f543e5c178e08b935eaf33bcc1d7ea23c030cfb716926fe5ca79788df814c77c: Status 404 returned error can't find the container with id f543e5c178e08b935eaf33bcc1d7ea23c030cfb716926fe5ca79788df814c77c Mar 17 09:29:53 crc kubenswrapper[4813]: I0317 09:29:53.955806 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twx4n\" (UniqueName: \"kubernetes.io/projected/35511914-5b8f-45e4-87cb-cd174e467dcc-kube-api-access-twx4n\") on node \"crc\" DevicePath \"\"" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.159429 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-etc-swift\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:29:54 crc kubenswrapper[4813]: E0317 09:29:54.159665 4813 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 17 09:29:54 crc kubenswrapper[4813]: E0317 09:29:54.159695 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 17 09:29:54 crc kubenswrapper[4813]: E0317 09:29:54.159767 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-etc-swift podName:76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51 nodeName:}" failed. No retries permitted until 2026-03-17 09:29:56.159746682 +0000 UTC m=+1218.260550191 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-etc-swift") pod "swift-storage-0" (UID: "76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51") : configmap "swift-ring-files" not found Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.354898 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-46tzl"] Mar 17 09:29:54 crc kubenswrapper[4813]: E0317 09:29:54.355298 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35511914-5b8f-45e4-87cb-cd174e467dcc" containerName="mariadb-account-create-update" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.355319 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="35511914-5b8f-45e4-87cb-cd174e467dcc" containerName="mariadb-account-create-update" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.355548 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="35511914-5b8f-45e4-87cb-cd174e467dcc" containerName="mariadb-account-create-update" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.356216 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-46tzl" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.368895 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-46tzl"] Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.459081 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-efef-account-create-update-wplm2"] Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.460243 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-efef-account-create-update-wplm2" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.463654 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.463688 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc9zt\" (UniqueName: \"kubernetes.io/projected/0f7359df-58d8-4c25-91b5-d262b088d46e-kube-api-access-pc9zt\") pod \"glance-db-create-46tzl\" (UID: \"0f7359df-58d8-4c25-91b5-d262b088d46e\") " pod="openstack/glance-db-create-46tzl" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.463815 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f7359df-58d8-4c25-91b5-d262b088d46e-operator-scripts\") pod \"glance-db-create-46tzl\" (UID: \"0f7359df-58d8-4c25-91b5-d262b088d46e\") " pod="openstack/glance-db-create-46tzl" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.474403 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-efef-account-create-update-wplm2"] Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.543391 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-rrfrc" event={"ID":"625a6e48-f069-49a9-b11a-342e12dffece","Type":"ContainerStarted","Data":"f543e5c178e08b935eaf33bcc1d7ea23c030cfb716926fe5ca79788df814c77c"} Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.545717 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6ee7-account-create-update-ffh6m" event={"ID":"35511914-5b8f-45e4-87cb-cd174e467dcc","Type":"ContainerDied","Data":"6c93c91cbe1e2fa43f2d31ddf6e1f46210bae5b23a4363e6ba9073adbf759542"} Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.545790 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c93c91cbe1e2fa43f2d31ddf6e1f46210bae5b23a4363e6ba9073adbf759542" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.545892 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6ee7-account-create-update-ffh6m" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.546299 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-wcmzb" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.576419 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f7359df-58d8-4c25-91b5-d262b088d46e-operator-scripts\") pod \"glance-db-create-46tzl\" (UID: \"0f7359df-58d8-4c25-91b5-d262b088d46e\") " pod="openstack/glance-db-create-46tzl" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.576547 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lqdks\" (UniqueName: \"kubernetes.io/projected/6b380124-90e1-48e8-a3f1-fca49c1adf07-kube-api-access-lqdks\") pod \"glance-efef-account-create-update-wplm2\" (UID: \"6b380124-90e1-48e8-a3f1-fca49c1adf07\") " pod="openstack/glance-efef-account-create-update-wplm2" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.577175 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc9zt\" (UniqueName: \"kubernetes.io/projected/0f7359df-58d8-4c25-91b5-d262b088d46e-kube-api-access-pc9zt\") pod \"glance-db-create-46tzl\" (UID: \"0f7359df-58d8-4c25-91b5-d262b088d46e\") " pod="openstack/glance-db-create-46tzl" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.577530 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f7359df-58d8-4c25-91b5-d262b088d46e-operator-scripts\") pod \"glance-db-create-46tzl\" (UID: \"0f7359df-58d8-4c25-91b5-d262b088d46e\") " pod="openstack/glance-db-create-46tzl" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.578369 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b380124-90e1-48e8-a3f1-fca49c1adf07-operator-scripts\") pod \"glance-efef-account-create-update-wplm2\" (UID: \"6b380124-90e1-48e8-a3f1-fca49c1adf07\") " pod="openstack/glance-efef-account-create-update-wplm2" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.611105 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc9zt\" (UniqueName: \"kubernetes.io/projected/0f7359df-58d8-4c25-91b5-d262b088d46e-kube-api-access-pc9zt\") pod \"glance-db-create-46tzl\" (UID: \"0f7359df-58d8-4c25-91b5-d262b088d46e\") " pod="openstack/glance-db-create-46tzl" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.626518 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/swift-ring-rebalance-wcmzb"] Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.635115 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/swift-ring-rebalance-wcmzb"] Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.677392 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-46tzl" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.681452 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lqdks\" (UniqueName: \"kubernetes.io/projected/6b380124-90e1-48e8-a3f1-fca49c1adf07-kube-api-access-lqdks\") pod \"glance-efef-account-create-update-wplm2\" (UID: \"6b380124-90e1-48e8-a3f1-fca49c1adf07\") " pod="openstack/glance-efef-account-create-update-wplm2" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.681625 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b380124-90e1-48e8-a3f1-fca49c1adf07-operator-scripts\") pod \"glance-efef-account-create-update-wplm2\" (UID: \"6b380124-90e1-48e8-a3f1-fca49c1adf07\") " pod="openstack/glance-efef-account-create-update-wplm2" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.682556 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b380124-90e1-48e8-a3f1-fca49c1adf07-operator-scripts\") pod \"glance-efef-account-create-update-wplm2\" (UID: \"6b380124-90e1-48e8-a3f1-fca49c1adf07\") " pod="openstack/glance-efef-account-create-update-wplm2" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.703968 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lqdks\" (UniqueName: \"kubernetes.io/projected/6b380124-90e1-48e8-a3f1-fca49c1adf07-kube-api-access-lqdks\") pod \"glance-efef-account-create-update-wplm2\" (UID: \"6b380124-90e1-48e8-a3f1-fca49c1adf07\") " pod="openstack/glance-efef-account-create-update-wplm2" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.751654 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e94939bc-6595-45af-94b8-148e850b86ff" path="/var/lib/kubelet/pods/e94939bc-6595-45af-94b8-148e850b86ff/volumes" Mar 17 09:29:54 crc kubenswrapper[4813]: I0317 09:29:54.774203 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-efef-account-create-update-wplm2" Mar 17 09:29:55 crc kubenswrapper[4813]: I0317 09:29:55.195887 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-46tzl"] Mar 17 09:29:55 crc kubenswrapper[4813]: W0317 09:29:55.201809 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f7359df_58d8_4c25_91b5_d262b088d46e.slice/crio-d7583fed1d59e4118544b5864bed30979daea5d4cf9962eee854e8548b5b5570 WatchSource:0}: Error finding container d7583fed1d59e4118544b5864bed30979daea5d4cf9962eee854e8548b5b5570: Status 404 returned error can't find the container with id d7583fed1d59e4118544b5864bed30979daea5d4cf9962eee854e8548b5b5570 Mar 17 09:29:55 crc kubenswrapper[4813]: I0317 09:29:55.291444 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-efef-account-create-update-wplm2"] Mar 17 09:29:55 crc kubenswrapper[4813]: W0317 09:29:55.293692 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6b380124_90e1_48e8_a3f1_fca49c1adf07.slice/crio-1c34ecdabe9ecb73a1ec33d60210224a44acbd696156d2c1abc0410f82734a7c WatchSource:0}: Error finding container 1c34ecdabe9ecb73a1ec33d60210224a44acbd696156d2c1abc0410f82734a7c: Status 404 returned error can't find the container with id 1c34ecdabe9ecb73a1ec33d60210224a44acbd696156d2c1abc0410f82734a7c Mar 17 09:29:55 crc kubenswrapper[4813]: I0317 09:29:55.556343 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-efef-account-create-update-wplm2" event={"ID":"6b380124-90e1-48e8-a3f1-fca49c1adf07","Type":"ContainerStarted","Data":"917eabd06202b2d1648e0f49f1227c65f4e00a2c5d4af77ea4e9b631ae4cfce5"} Mar 17 09:29:55 crc kubenswrapper[4813]: I0317 09:29:55.556655 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-efef-account-create-update-wplm2" event={"ID":"6b380124-90e1-48e8-a3f1-fca49c1adf07","Type":"ContainerStarted","Data":"1c34ecdabe9ecb73a1ec33d60210224a44acbd696156d2c1abc0410f82734a7c"} Mar 17 09:29:55 crc kubenswrapper[4813]: I0317 09:29:55.557917 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-46tzl" event={"ID":"0f7359df-58d8-4c25-91b5-d262b088d46e","Type":"ContainerStarted","Data":"77f9b0dc665da47be2b72e324542837824014ca8377b7d1a66110ff7f2ba8f38"} Mar 17 09:29:55 crc kubenswrapper[4813]: I0317 09:29:55.557945 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-46tzl" event={"ID":"0f7359df-58d8-4c25-91b5-d262b088d46e","Type":"ContainerStarted","Data":"d7583fed1d59e4118544b5864bed30979daea5d4cf9962eee854e8548b5b5570"} Mar 17 09:29:55 crc kubenswrapper[4813]: I0317 09:29:55.580458 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-efef-account-create-update-wplm2" podStartSLOduration=1.580439583 podStartE2EDuration="1.580439583s" podCreationTimestamp="2026-03-17 09:29:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:29:55.567670302 +0000 UTC m=+1217.668473811" watchObservedRunningTime="2026-03-17 09:29:55.580439583 +0000 UTC m=+1217.681243072" Mar 17 09:29:55 crc kubenswrapper[4813]: I0317 09:29:55.586457 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-create-46tzl" podStartSLOduration=1.586437042 podStartE2EDuration="1.586437042s" podCreationTimestamp="2026-03-17 09:29:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:29:55.585803362 +0000 UTC m=+1217.686606861" watchObservedRunningTime="2026-03-17 09:29:55.586437042 +0000 UTC m=+1217.687240541" Mar 17 09:29:56 crc kubenswrapper[4813]: I0317 09:29:56.029588 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-8sdrx"] Mar 17 09:29:56 crc kubenswrapper[4813]: I0317 09:29:56.038532 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-8sdrx" Mar 17 09:29:56 crc kubenswrapper[4813]: I0317 09:29:56.041207 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 17 09:29:56 crc kubenswrapper[4813]: I0317 09:29:56.071321 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-8sdrx"] Mar 17 09:29:56 crc kubenswrapper[4813]: I0317 09:29:56.110178 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7gkq\" (UniqueName: \"kubernetes.io/projected/4296c965-29b4-4676-80ba-959682f0d867-kube-api-access-j7gkq\") pod \"root-account-create-update-8sdrx\" (UID: \"4296c965-29b4-4676-80ba-959682f0d867\") " pod="openstack/root-account-create-update-8sdrx" Mar 17 09:29:56 crc kubenswrapper[4813]: I0317 09:29:56.110271 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4296c965-29b4-4676-80ba-959682f0d867-operator-scripts\") pod \"root-account-create-update-8sdrx\" (UID: \"4296c965-29b4-4676-80ba-959682f0d867\") " pod="openstack/root-account-create-update-8sdrx" Mar 17 09:29:56 crc kubenswrapper[4813]: I0317 09:29:56.212380 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4296c965-29b4-4676-80ba-959682f0d867-operator-scripts\") pod \"root-account-create-update-8sdrx\" (UID: \"4296c965-29b4-4676-80ba-959682f0d867\") " pod="openstack/root-account-create-update-8sdrx" Mar 17 09:29:56 crc kubenswrapper[4813]: I0317 09:29:56.212874 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-etc-swift\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:29:56 crc kubenswrapper[4813]: I0317 09:29:56.213029 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7gkq\" (UniqueName: \"kubernetes.io/projected/4296c965-29b4-4676-80ba-959682f0d867-kube-api-access-j7gkq\") pod \"root-account-create-update-8sdrx\" (UID: \"4296c965-29b4-4676-80ba-959682f0d867\") " pod="openstack/root-account-create-update-8sdrx" Mar 17 09:29:56 crc kubenswrapper[4813]: E0317 09:29:56.213088 4813 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 17 09:29:56 crc kubenswrapper[4813]: E0317 09:29:56.213116 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 17 09:29:56 crc kubenswrapper[4813]: E0317 09:29:56.213186 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-etc-swift podName:76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51 nodeName:}" failed. No retries permitted until 2026-03-17 09:30:00.213169243 +0000 UTC m=+1222.313972742 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-etc-swift") pod "swift-storage-0" (UID: "76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51") : configmap "swift-ring-files" not found Mar 17 09:29:56 crc kubenswrapper[4813]: I0317 09:29:56.213845 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4296c965-29b4-4676-80ba-959682f0d867-operator-scripts\") pod \"root-account-create-update-8sdrx\" (UID: \"4296c965-29b4-4676-80ba-959682f0d867\") " pod="openstack/root-account-create-update-8sdrx" Mar 17 09:29:56 crc kubenswrapper[4813]: I0317 09:29:56.245806 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7gkq\" (UniqueName: \"kubernetes.io/projected/4296c965-29b4-4676-80ba-959682f0d867-kube-api-access-j7gkq\") pod \"root-account-create-update-8sdrx\" (UID: \"4296c965-29b4-4676-80ba-959682f0d867\") " pod="openstack/root-account-create-update-8sdrx" Mar 17 09:29:56 crc kubenswrapper[4813]: I0317 09:29:56.370581 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-8sdrx" Mar 17 09:29:56 crc kubenswrapper[4813]: I0317 09:29:56.573828 4813 generic.go:334] "Generic (PLEG): container finished" podID="6b380124-90e1-48e8-a3f1-fca49c1adf07" containerID="917eabd06202b2d1648e0f49f1227c65f4e00a2c5d4af77ea4e9b631ae4cfce5" exitCode=0 Mar 17 09:29:56 crc kubenswrapper[4813]: I0317 09:29:56.573960 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-efef-account-create-update-wplm2" event={"ID":"6b380124-90e1-48e8-a3f1-fca49c1adf07","Type":"ContainerDied","Data":"917eabd06202b2d1648e0f49f1227c65f4e00a2c5d4af77ea4e9b631ae4cfce5"} Mar 17 09:29:56 crc kubenswrapper[4813]: I0317 09:29:56.587807 4813 generic.go:334] "Generic (PLEG): container finished" podID="0f7359df-58d8-4c25-91b5-d262b088d46e" containerID="77f9b0dc665da47be2b72e324542837824014ca8377b7d1a66110ff7f2ba8f38" exitCode=0 Mar 17 09:29:56 crc kubenswrapper[4813]: I0317 09:29:56.587858 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-46tzl" event={"ID":"0f7359df-58d8-4c25-91b5-d262b088d46e","Type":"ContainerDied","Data":"77f9b0dc665da47be2b72e324542837824014ca8377b7d1a66110ff7f2ba8f38"} Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.161071 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562330-b5zg5"] Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.162637 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562330-b5zg5" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.165755 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.166406 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.166716 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.186806 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562330-qjkx6"] Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.188103 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562330-qjkx6" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.191039 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.191095 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.214261 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562330-qjkx6"] Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.223347 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562330-b5zg5"] Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.235485 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-s6zhg"] Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.236893 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-s6zhg" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.261689 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-s6zhg"] Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.289025 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-etc-swift\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.289079 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/47bda8a6-acb6-48df-b15d-1561c26df6cf-secret-volume\") pod \"collect-profiles-29562330-qjkx6\" (UID: \"47bda8a6-acb6-48df-b15d-1561c26df6cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562330-qjkx6" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.289109 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/47bda8a6-acb6-48df-b15d-1561c26df6cf-config-volume\") pod \"collect-profiles-29562330-qjkx6\" (UID: \"47bda8a6-acb6-48df-b15d-1561c26df6cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562330-qjkx6" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.289161 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqzkj\" (UniqueName: \"kubernetes.io/projected/0cbf9e10-aa11-4c2e-9930-b4f1f451bc65-kube-api-access-nqzkj\") pod \"keystone-db-create-s6zhg\" (UID: \"0cbf9e10-aa11-4c2e-9930-b4f1f451bc65\") " pod="openstack/keystone-db-create-s6zhg" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.289192 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cg7cv\" (UniqueName: \"kubernetes.io/projected/339d3b99-3aa5-4c3b-99b2-997de272c26c-kube-api-access-cg7cv\") pod \"auto-csr-approver-29562330-b5zg5\" (UID: \"339d3b99-3aa5-4c3b-99b2-997de272c26c\") " pod="openshift-infra/auto-csr-approver-29562330-b5zg5" Mar 17 09:30:00 crc kubenswrapper[4813]: E0317 09:30:00.289210 4813 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.289235 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vzfbn\" (UniqueName: \"kubernetes.io/projected/47bda8a6-acb6-48df-b15d-1561c26df6cf-kube-api-access-vzfbn\") pod \"collect-profiles-29562330-qjkx6\" (UID: \"47bda8a6-acb6-48df-b15d-1561c26df6cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562330-qjkx6" Mar 17 09:30:00 crc kubenswrapper[4813]: E0317 09:30:00.289243 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.289252 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0cbf9e10-aa11-4c2e-9930-b4f1f451bc65-operator-scripts\") pod \"keystone-db-create-s6zhg\" (UID: \"0cbf9e10-aa11-4c2e-9930-b4f1f451bc65\") " pod="openstack/keystone-db-create-s6zhg" Mar 17 09:30:00 crc kubenswrapper[4813]: E0317 09:30:00.289299 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-etc-swift podName:76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51 nodeName:}" failed. No retries permitted until 2026-03-17 09:30:08.28928072 +0000 UTC m=+1230.390084299 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-etc-swift") pod "swift-storage-0" (UID: "76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51") : configmap "swift-ring-files" not found Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.309859 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-2d07-account-create-update-4j8f9"] Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.311853 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2d07-account-create-update-4j8f9" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.314831 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.323467 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2d07-account-create-update-4j8f9"] Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.390654 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cg7cv\" (UniqueName: \"kubernetes.io/projected/339d3b99-3aa5-4c3b-99b2-997de272c26c-kube-api-access-cg7cv\") pod \"auto-csr-approver-29562330-b5zg5\" (UID: \"339d3b99-3aa5-4c3b-99b2-997de272c26c\") " pod="openshift-infra/auto-csr-approver-29562330-b5zg5" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.390783 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vzfbn\" (UniqueName: \"kubernetes.io/projected/47bda8a6-acb6-48df-b15d-1561c26df6cf-kube-api-access-vzfbn\") pod \"collect-profiles-29562330-qjkx6\" (UID: \"47bda8a6-acb6-48df-b15d-1561c26df6cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562330-qjkx6" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.390812 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0cbf9e10-aa11-4c2e-9930-b4f1f451bc65-operator-scripts\") pod \"keystone-db-create-s6zhg\" (UID: \"0cbf9e10-aa11-4c2e-9930-b4f1f451bc65\") " pod="openstack/keystone-db-create-s6zhg" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.391189 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjwtp\" (UniqueName: \"kubernetes.io/projected/e4b19fd6-39e5-49da-811a-81b0ce1f0c29-kube-api-access-kjwtp\") pod \"keystone-2d07-account-create-update-4j8f9\" (UID: \"e4b19fd6-39e5-49da-811a-81b0ce1f0c29\") " pod="openstack/keystone-2d07-account-create-update-4j8f9" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.391234 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/47bda8a6-acb6-48df-b15d-1561c26df6cf-secret-volume\") pod \"collect-profiles-29562330-qjkx6\" (UID: \"47bda8a6-acb6-48df-b15d-1561c26df6cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562330-qjkx6" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.391775 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0cbf9e10-aa11-4c2e-9930-b4f1f451bc65-operator-scripts\") pod \"keystone-db-create-s6zhg\" (UID: \"0cbf9e10-aa11-4c2e-9930-b4f1f451bc65\") " pod="openstack/keystone-db-create-s6zhg" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.392083 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/47bda8a6-acb6-48df-b15d-1561c26df6cf-config-volume\") pod \"collect-profiles-29562330-qjkx6\" (UID: \"47bda8a6-acb6-48df-b15d-1561c26df6cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562330-qjkx6" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.392127 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4b19fd6-39e5-49da-811a-81b0ce1f0c29-operator-scripts\") pod \"keystone-2d07-account-create-update-4j8f9\" (UID: \"e4b19fd6-39e5-49da-811a-81b0ce1f0c29\") " pod="openstack/keystone-2d07-account-create-update-4j8f9" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.392206 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqzkj\" (UniqueName: \"kubernetes.io/projected/0cbf9e10-aa11-4c2e-9930-b4f1f451bc65-kube-api-access-nqzkj\") pod \"keystone-db-create-s6zhg\" (UID: \"0cbf9e10-aa11-4c2e-9930-b4f1f451bc65\") " pod="openstack/keystone-db-create-s6zhg" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.392332 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/47bda8a6-acb6-48df-b15d-1561c26df6cf-config-volume\") pod \"collect-profiles-29562330-qjkx6\" (UID: \"47bda8a6-acb6-48df-b15d-1561c26df6cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562330-qjkx6" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.407763 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/47bda8a6-acb6-48df-b15d-1561c26df6cf-secret-volume\") pod \"collect-profiles-29562330-qjkx6\" (UID: \"47bda8a6-acb6-48df-b15d-1561c26df6cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562330-qjkx6" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.415323 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqzkj\" (UniqueName: \"kubernetes.io/projected/0cbf9e10-aa11-4c2e-9930-b4f1f451bc65-kube-api-access-nqzkj\") pod \"keystone-db-create-s6zhg\" (UID: \"0cbf9e10-aa11-4c2e-9930-b4f1f451bc65\") " pod="openstack/keystone-db-create-s6zhg" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.416563 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cg7cv\" (UniqueName: \"kubernetes.io/projected/339d3b99-3aa5-4c3b-99b2-997de272c26c-kube-api-access-cg7cv\") pod \"auto-csr-approver-29562330-b5zg5\" (UID: \"339d3b99-3aa5-4c3b-99b2-997de272c26c\") " pod="openshift-infra/auto-csr-approver-29562330-b5zg5" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.416690 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vzfbn\" (UniqueName: \"kubernetes.io/projected/47bda8a6-acb6-48df-b15d-1561c26df6cf-kube-api-access-vzfbn\") pod \"collect-profiles-29562330-qjkx6\" (UID: \"47bda8a6-acb6-48df-b15d-1561c26df6cf\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562330-qjkx6" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.496062 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjwtp\" (UniqueName: \"kubernetes.io/projected/e4b19fd6-39e5-49da-811a-81b0ce1f0c29-kube-api-access-kjwtp\") pod \"keystone-2d07-account-create-update-4j8f9\" (UID: \"e4b19fd6-39e5-49da-811a-81b0ce1f0c29\") " pod="openstack/keystone-2d07-account-create-update-4j8f9" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.496423 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4b19fd6-39e5-49da-811a-81b0ce1f0c29-operator-scripts\") pod \"keystone-2d07-account-create-update-4j8f9\" (UID: \"e4b19fd6-39e5-49da-811a-81b0ce1f0c29\") " pod="openstack/keystone-2d07-account-create-update-4j8f9" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.497318 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4b19fd6-39e5-49da-811a-81b0ce1f0c29-operator-scripts\") pod \"keystone-2d07-account-create-update-4j8f9\" (UID: \"e4b19fd6-39e5-49da-811a-81b0ce1f0c29\") " pod="openstack/keystone-2d07-account-create-update-4j8f9" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.502263 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-8w5d7"] Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.503398 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-8w5d7" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.510287 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562330-b5zg5" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.511903 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-8w5d7"] Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.518014 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562330-qjkx6" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.519134 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjwtp\" (UniqueName: \"kubernetes.io/projected/e4b19fd6-39e5-49da-811a-81b0ce1f0c29-kube-api-access-kjwtp\") pod \"keystone-2d07-account-create-update-4j8f9\" (UID: \"e4b19fd6-39e5-49da-811a-81b0ce1f0c29\") " pod="openstack/keystone-2d07-account-create-update-4j8f9" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.553689 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-s6zhg" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.597798 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgw4x\" (UniqueName: \"kubernetes.io/projected/d7b2d571-5a93-4520-8bab-324e4d27bdc9-kube-api-access-cgw4x\") pod \"placement-db-create-8w5d7\" (UID: \"d7b2d571-5a93-4520-8bab-324e4d27bdc9\") " pod="openstack/placement-db-create-8w5d7" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.597884 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b2d571-5a93-4520-8bab-324e4d27bdc9-operator-scripts\") pod \"placement-db-create-8w5d7\" (UID: \"d7b2d571-5a93-4520-8bab-324e4d27bdc9\") " pod="openstack/placement-db-create-8w5d7" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.631748 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2d07-account-create-update-4j8f9" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.699681 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgw4x\" (UniqueName: \"kubernetes.io/projected/d7b2d571-5a93-4520-8bab-324e4d27bdc9-kube-api-access-cgw4x\") pod \"placement-db-create-8w5d7\" (UID: \"d7b2d571-5a93-4520-8bab-324e4d27bdc9\") " pod="openstack/placement-db-create-8w5d7" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.699786 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b2d571-5a93-4520-8bab-324e4d27bdc9-operator-scripts\") pod \"placement-db-create-8w5d7\" (UID: \"d7b2d571-5a93-4520-8bab-324e4d27bdc9\") " pod="openstack/placement-db-create-8w5d7" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.702116 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b2d571-5a93-4520-8bab-324e4d27bdc9-operator-scripts\") pod \"placement-db-create-8w5d7\" (UID: \"d7b2d571-5a93-4520-8bab-324e4d27bdc9\") " pod="openstack/placement-db-create-8w5d7" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.716882 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgw4x\" (UniqueName: \"kubernetes.io/projected/d7b2d571-5a93-4520-8bab-324e4d27bdc9-kube-api-access-cgw4x\") pod \"placement-db-create-8w5d7\" (UID: \"d7b2d571-5a93-4520-8bab-324e4d27bdc9\") " pod="openstack/placement-db-create-8w5d7" Mar 17 09:30:00 crc kubenswrapper[4813]: I0317 09:30:00.851412 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-8w5d7" Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.381901 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-46tzl" Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.388943 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-efef-account-create-update-wplm2" Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.412121 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b380124-90e1-48e8-a3f1-fca49c1adf07-operator-scripts\") pod \"6b380124-90e1-48e8-a3f1-fca49c1adf07\" (UID: \"6b380124-90e1-48e8-a3f1-fca49c1adf07\") " Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.412342 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lqdks\" (UniqueName: \"kubernetes.io/projected/6b380124-90e1-48e8-a3f1-fca49c1adf07-kube-api-access-lqdks\") pod \"6b380124-90e1-48e8-a3f1-fca49c1adf07\" (UID: \"6b380124-90e1-48e8-a3f1-fca49c1adf07\") " Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.412465 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pc9zt\" (UniqueName: \"kubernetes.io/projected/0f7359df-58d8-4c25-91b5-d262b088d46e-kube-api-access-pc9zt\") pod \"0f7359df-58d8-4c25-91b5-d262b088d46e\" (UID: \"0f7359df-58d8-4c25-91b5-d262b088d46e\") " Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.412511 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f7359df-58d8-4c25-91b5-d262b088d46e-operator-scripts\") pod \"0f7359df-58d8-4c25-91b5-d262b088d46e\" (UID: \"0f7359df-58d8-4c25-91b5-d262b088d46e\") " Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.413269 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b380124-90e1-48e8-a3f1-fca49c1adf07-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6b380124-90e1-48e8-a3f1-fca49c1adf07" (UID: "6b380124-90e1-48e8-a3f1-fca49c1adf07"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.413793 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0f7359df-58d8-4c25-91b5-d262b088d46e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0f7359df-58d8-4c25-91b5-d262b088d46e" (UID: "0f7359df-58d8-4c25-91b5-d262b088d46e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.417474 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b380124-90e1-48e8-a3f1-fca49c1adf07-kube-api-access-lqdks" (OuterVolumeSpecName: "kube-api-access-lqdks") pod "6b380124-90e1-48e8-a3f1-fca49c1adf07" (UID: "6b380124-90e1-48e8-a3f1-fca49c1adf07"). InnerVolumeSpecName "kube-api-access-lqdks". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.421787 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f7359df-58d8-4c25-91b5-d262b088d46e-kube-api-access-pc9zt" (OuterVolumeSpecName: "kube-api-access-pc9zt") pod "0f7359df-58d8-4c25-91b5-d262b088d46e" (UID: "0f7359df-58d8-4c25-91b5-d262b088d46e"). InnerVolumeSpecName "kube-api-access-pc9zt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.518294 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pc9zt\" (UniqueName: \"kubernetes.io/projected/0f7359df-58d8-4c25-91b5-d262b088d46e-kube-api-access-pc9zt\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.518332 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0f7359df-58d8-4c25-91b5-d262b088d46e-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.518346 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6b380124-90e1-48e8-a3f1-fca49c1adf07-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.518375 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lqdks\" (UniqueName: \"kubernetes.io/projected/6b380124-90e1-48e8-a3f1-fca49c1adf07-kube-api-access-lqdks\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.632994 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-46tzl" Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.632992 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-46tzl" event={"ID":"0f7359df-58d8-4c25-91b5-d262b088d46e","Type":"ContainerDied","Data":"d7583fed1d59e4118544b5864bed30979daea5d4cf9962eee854e8548b5b5570"} Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.633256 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7583fed1d59e4118544b5864bed30979daea5d4cf9962eee854e8548b5b5570" Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.636238 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-efef-account-create-update-wplm2" Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.636169 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-efef-account-create-update-wplm2" event={"ID":"6b380124-90e1-48e8-a3f1-fca49c1adf07","Type":"ContainerDied","Data":"1c34ecdabe9ecb73a1ec33d60210224a44acbd696156d2c1abc0410f82734a7c"} Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.636414 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c34ecdabe9ecb73a1ec33d60210224a44acbd696156d2c1abc0410f82734a7c" Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.645692 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-b25vw" Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.700297 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-rn42g"] Mar 17 09:30:01 crc kubenswrapper[4813]: I0317 09:30:01.701375 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" podUID="7f68b346-4221-4e2f-bb64-c43040aeb3cd" containerName="dnsmasq-dns" containerID="cri-o://28d1246593453fd59a62975aeeefcc594ae26fe13729dc3b45ce495a59cb60bc" gracePeriod=10 Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.222814 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-2d07-account-create-update-4j8f9"] Mar 17 09:30:02 crc kubenswrapper[4813]: W0317 09:30:02.245109 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4b19fd6_39e5_49da_811a_81b0ce1f0c29.slice/crio-366073ccc2c12696b643495e39cf29ec904f5f9813f393dc88491c56fd863885 WatchSource:0}: Error finding container 366073ccc2c12696b643495e39cf29ec904f5f9813f393dc88491c56fd863885: Status 404 returned error can't find the container with id 366073ccc2c12696b643495e39cf29ec904f5f9813f393dc88491c56fd863885 Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.355160 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562330-qjkx6"] Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.360657 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.363488 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562330-b5zg5"] Mar 17 09:30:02 crc kubenswrapper[4813]: W0317 09:30:02.368319 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod47bda8a6_acb6_48df_b15d_1561c26df6cf.slice/crio-dbb9a3412b7ea1fb0d5e822c5499ee8e31f9a4750c4e89eb5b7ae8536b20f402 WatchSource:0}: Error finding container dbb9a3412b7ea1fb0d5e822c5499ee8e31f9a4750c4e89eb5b7ae8536b20f402: Status 404 returned error can't find the container with id dbb9a3412b7ea1fb0d5e822c5499ee8e31f9a4750c4e89eb5b7ae8536b20f402 Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.371501 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-s6zhg"] Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.392469 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-8sdrx"] Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.427890 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-8w5d7"] Mar 17 09:30:02 crc kubenswrapper[4813]: W0317 09:30:02.433724 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7b2d571_5a93_4520_8bab_324e4d27bdc9.slice/crio-450ac5009ea4d1336b657eda866aadf5d919fd253fd9d4d7d5f3d39a8bae4ceb WatchSource:0}: Error finding container 450ac5009ea4d1336b657eda866aadf5d919fd253fd9d4d7d5f3d39a8bae4ceb: Status 404 returned error can't find the container with id 450ac5009ea4d1336b657eda866aadf5d919fd253fd9d4d7d5f3d39a8bae4ceb Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.440432 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-dns-svc\") pod \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\" (UID: \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\") " Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.440472 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-ovsdbserver-nb\") pod \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\" (UID: \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\") " Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.440532 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-ovsdbserver-sb\") pod \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\" (UID: \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\") " Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.440653 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-config\") pod \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\" (UID: \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\") " Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.440714 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pn7nr\" (UniqueName: \"kubernetes.io/projected/7f68b346-4221-4e2f-bb64-c43040aeb3cd-kube-api-access-pn7nr\") pod \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\" (UID: \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\") " Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.474761 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f68b346-4221-4e2f-bb64-c43040aeb3cd-kube-api-access-pn7nr" (OuterVolumeSpecName: "kube-api-access-pn7nr") pod "7f68b346-4221-4e2f-bb64-c43040aeb3cd" (UID: "7f68b346-4221-4e2f-bb64-c43040aeb3cd"). InnerVolumeSpecName "kube-api-access-pn7nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.516868 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7f68b346-4221-4e2f-bb64-c43040aeb3cd" (UID: "7f68b346-4221-4e2f-bb64-c43040aeb3cd"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.522981 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7f68b346-4221-4e2f-bb64-c43040aeb3cd" (UID: "7f68b346-4221-4e2f-bb64-c43040aeb3cd"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.538369 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-config" (OuterVolumeSpecName: "config") pod "7f68b346-4221-4e2f-bb64-c43040aeb3cd" (UID: "7f68b346-4221-4e2f-bb64-c43040aeb3cd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.543827 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7f68b346-4221-4e2f-bb64-c43040aeb3cd" (UID: "7f68b346-4221-4e2f-bb64-c43040aeb3cd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.545079 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-ovsdbserver-nb\") pod \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\" (UID: \"7f68b346-4221-4e2f-bb64-c43040aeb3cd\") " Mar 17 09:30:02 crc kubenswrapper[4813]: W0317 09:30:02.546245 4813 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/7f68b346-4221-4e2f-bb64-c43040aeb3cd/volumes/kubernetes.io~configmap/ovsdbserver-nb Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.546343 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7f68b346-4221-4e2f-bb64-c43040aeb3cd" (UID: "7f68b346-4221-4e2f-bb64-c43040aeb3cd"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.547308 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.547425 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pn7nr\" (UniqueName: \"kubernetes.io/projected/7f68b346-4221-4e2f-bb64-c43040aeb3cd-kube-api-access-pn7nr\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.547530 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.547629 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.547717 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f68b346-4221-4e2f-bb64-c43040aeb3cd-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.647131 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-8w5d7" event={"ID":"d7b2d571-5a93-4520-8bab-324e4d27bdc9","Type":"ContainerStarted","Data":"450ac5009ea4d1336b657eda866aadf5d919fd253fd9d4d7d5f3d39a8bae4ceb"} Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.649746 4813 generic.go:334] "Generic (PLEG): container finished" podID="7f68b346-4221-4e2f-bb64-c43040aeb3cd" containerID="28d1246593453fd59a62975aeeefcc594ae26fe13729dc3b45ce495a59cb60bc" exitCode=0 Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.649849 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" event={"ID":"7f68b346-4221-4e2f-bb64-c43040aeb3cd","Type":"ContainerDied","Data":"28d1246593453fd59a62975aeeefcc594ae26fe13729dc3b45ce495a59cb60bc"} Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.649889 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" event={"ID":"7f68b346-4221-4e2f-bb64-c43040aeb3cd","Type":"ContainerDied","Data":"c230733c0c33ba3d2325c3ce1a27b75dd746d99f37079ff99315d3184e2a7e08"} Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.649913 4813 scope.go:117] "RemoveContainer" containerID="28d1246593453fd59a62975aeeefcc594ae26fe13729dc3b45ce495a59cb60bc" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.650075 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-86db49b7ff-rn42g" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.658411 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562330-b5zg5" event={"ID":"339d3b99-3aa5-4c3b-99b2-997de272c26c","Type":"ContainerStarted","Data":"776c488601eb81dbe228ff8b21929c9a559c1cdba9b8e110794278329856ee94"} Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.663166 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2d07-account-create-update-4j8f9" event={"ID":"e4b19fd6-39e5-49da-811a-81b0ce1f0c29","Type":"ContainerStarted","Data":"7de276f53292a53bf05b1f8194df75b83e957f1cb6a3b52b9cd3f27bb0f7c3d3"} Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.663218 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2d07-account-create-update-4j8f9" event={"ID":"e4b19fd6-39e5-49da-811a-81b0ce1f0c29","Type":"ContainerStarted","Data":"366073ccc2c12696b643495e39cf29ec904f5f9813f393dc88491c56fd863885"} Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.666696 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-rrfrc" event={"ID":"625a6e48-f069-49a9-b11a-342e12dffece","Type":"ContainerStarted","Data":"aa8f414377f9f29c12c6d9c5723ef631c4ea0afda6b0e5363d36678886bdcd01"} Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.667390 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-s6zhg" event={"ID":"0cbf9e10-aa11-4c2e-9930-b4f1f451bc65","Type":"ContainerStarted","Data":"2d536fd359c6cbcdbcefe8f7ee979ac807ebd5b093342748d3639a6cf64818ee"} Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.668419 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562330-qjkx6" event={"ID":"47bda8a6-acb6-48df-b15d-1561c26df6cf","Type":"ContainerStarted","Data":"dbb9a3412b7ea1fb0d5e822c5499ee8e31f9a4750c4e89eb5b7ae8536b20f402"} Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.670421 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-8sdrx" event={"ID":"4296c965-29b4-4676-80ba-959682f0d867","Type":"ContainerStarted","Data":"c997f66df63d205fdef0da50a5a4258eae986e0babbf6ddfe058c7c5ccd57c2e"} Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.675783 4813 scope.go:117] "RemoveContainer" containerID="0511cc02d9d271cdc5e9ac7bab910f26bbf0503afb30b260655b65979ce18f31" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.694778 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-2d07-account-create-update-4j8f9" podStartSLOduration=2.694754004 podStartE2EDuration="2.694754004s" podCreationTimestamp="2026-03-17 09:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:30:02.682573271 +0000 UTC m=+1224.783376770" watchObservedRunningTime="2026-03-17 09:30:02.694754004 +0000 UTC m=+1224.795557513" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.702556 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-rn42g"] Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.708792 4813 scope.go:117] "RemoveContainer" containerID="28d1246593453fd59a62975aeeefcc594ae26fe13729dc3b45ce495a59cb60bc" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.712915 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-86db49b7ff-rn42g"] Mar 17 09:30:02 crc kubenswrapper[4813]: E0317 09:30:02.714443 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"28d1246593453fd59a62975aeeefcc594ae26fe13729dc3b45ce495a59cb60bc\": container with ID starting with 28d1246593453fd59a62975aeeefcc594ae26fe13729dc3b45ce495a59cb60bc not found: ID does not exist" containerID="28d1246593453fd59a62975aeeefcc594ae26fe13729dc3b45ce495a59cb60bc" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.714492 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"28d1246593453fd59a62975aeeefcc594ae26fe13729dc3b45ce495a59cb60bc"} err="failed to get container status \"28d1246593453fd59a62975aeeefcc594ae26fe13729dc3b45ce495a59cb60bc\": rpc error: code = NotFound desc = could not find container \"28d1246593453fd59a62975aeeefcc594ae26fe13729dc3b45ce495a59cb60bc\": container with ID starting with 28d1246593453fd59a62975aeeefcc594ae26fe13729dc3b45ce495a59cb60bc not found: ID does not exist" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.714569 4813 scope.go:117] "RemoveContainer" containerID="0511cc02d9d271cdc5e9ac7bab910f26bbf0503afb30b260655b65979ce18f31" Mar 17 09:30:02 crc kubenswrapper[4813]: E0317 09:30:02.715014 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0511cc02d9d271cdc5e9ac7bab910f26bbf0503afb30b260655b65979ce18f31\": container with ID starting with 0511cc02d9d271cdc5e9ac7bab910f26bbf0503afb30b260655b65979ce18f31 not found: ID does not exist" containerID="0511cc02d9d271cdc5e9ac7bab910f26bbf0503afb30b260655b65979ce18f31" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.715056 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0511cc02d9d271cdc5e9ac7bab910f26bbf0503afb30b260655b65979ce18f31"} err="failed to get container status \"0511cc02d9d271cdc5e9ac7bab910f26bbf0503afb30b260655b65979ce18f31\": rpc error: code = NotFound desc = could not find container \"0511cc02d9d271cdc5e9ac7bab910f26bbf0503afb30b260655b65979ce18f31\": container with ID starting with 0511cc02d9d271cdc5e9ac7bab910f26bbf0503afb30b260655b65979ce18f31 not found: ID does not exist" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.726035 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-rrfrc" podStartSLOduration=2.088307067 podStartE2EDuration="9.726013888s" podCreationTimestamp="2026-03-17 09:29:53 +0000 UTC" firstStartedPulling="2026-03-17 09:29:53.875965577 +0000 UTC m=+1215.976769076" lastFinishedPulling="2026-03-17 09:30:01.513672358 +0000 UTC m=+1223.614475897" observedRunningTime="2026-03-17 09:30:02.709182108 +0000 UTC m=+1224.809985607" watchObservedRunningTime="2026-03-17 09:30:02.726013888 +0000 UTC m=+1224.826817377" Mar 17 09:30:02 crc kubenswrapper[4813]: I0317 09:30:02.741490 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f68b346-4221-4e2f-bb64-c43040aeb3cd" path="/var/lib/kubelet/pods/7f68b346-4221-4e2f-bb64-c43040aeb3cd/volumes" Mar 17 09:30:03 crc kubenswrapper[4813]: I0317 09:30:03.682080 4813 generic.go:334] "Generic (PLEG): container finished" podID="0cbf9e10-aa11-4c2e-9930-b4f1f451bc65" containerID="b29e9b4eda5417c83566c4aafd4a22ca559e43d575dcc2c9328f0e67d69d820c" exitCode=0 Mar 17 09:30:03 crc kubenswrapper[4813]: I0317 09:30:03.682568 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-s6zhg" event={"ID":"0cbf9e10-aa11-4c2e-9930-b4f1f451bc65","Type":"ContainerDied","Data":"b29e9b4eda5417c83566c4aafd4a22ca559e43d575dcc2c9328f0e67d69d820c"} Mar 17 09:30:03 crc kubenswrapper[4813]: I0317 09:30:03.686654 4813 generic.go:334] "Generic (PLEG): container finished" podID="47bda8a6-acb6-48df-b15d-1561c26df6cf" containerID="f2d9102e40557554cd79652b1df0de5876a52bf58b1793baf2ffc6a404a7548b" exitCode=0 Mar 17 09:30:03 crc kubenswrapper[4813]: I0317 09:30:03.686803 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562330-qjkx6" event={"ID":"47bda8a6-acb6-48df-b15d-1561c26df6cf","Type":"ContainerDied","Data":"f2d9102e40557554cd79652b1df0de5876a52bf58b1793baf2ffc6a404a7548b"} Mar 17 09:30:03 crc kubenswrapper[4813]: I0317 09:30:03.689438 4813 generic.go:334] "Generic (PLEG): container finished" podID="4296c965-29b4-4676-80ba-959682f0d867" containerID="4447f4ba14b78f065ba1ae8a1456f1f1fd8401676d81db53dfe718ae82ef145d" exitCode=0 Mar 17 09:30:03 crc kubenswrapper[4813]: I0317 09:30:03.689500 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-8sdrx" event={"ID":"4296c965-29b4-4676-80ba-959682f0d867","Type":"ContainerDied","Data":"4447f4ba14b78f065ba1ae8a1456f1f1fd8401676d81db53dfe718ae82ef145d"} Mar 17 09:30:03 crc kubenswrapper[4813]: I0317 09:30:03.691185 4813 generic.go:334] "Generic (PLEG): container finished" podID="d7b2d571-5a93-4520-8bab-324e4d27bdc9" containerID="99c90a851496fedf51f09757cf22a17306cec4664540c90f864f16dc379cec24" exitCode=0 Mar 17 09:30:03 crc kubenswrapper[4813]: I0317 09:30:03.691240 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-8w5d7" event={"ID":"d7b2d571-5a93-4520-8bab-324e4d27bdc9","Type":"ContainerDied","Data":"99c90a851496fedf51f09757cf22a17306cec4664540c90f864f16dc379cec24"} Mar 17 09:30:03 crc kubenswrapper[4813]: I0317 09:30:03.694023 4813 generic.go:334] "Generic (PLEG): container finished" podID="e4b19fd6-39e5-49da-811a-81b0ce1f0c29" containerID="7de276f53292a53bf05b1f8194df75b83e957f1cb6a3b52b9cd3f27bb0f7c3d3" exitCode=0 Mar 17 09:30:03 crc kubenswrapper[4813]: I0317 09:30:03.694109 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2d07-account-create-update-4j8f9" event={"ID":"e4b19fd6-39e5-49da-811a-81b0ce1f0c29","Type":"ContainerDied","Data":"7de276f53292a53bf05b1f8194df75b83e957f1cb6a3b52b9cd3f27bb0f7c3d3"} Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.786365 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-nrh2h"] Mar 17 09:30:04 crc kubenswrapper[4813]: E0317 09:30:04.787231 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f68b346-4221-4e2f-bb64-c43040aeb3cd" containerName="dnsmasq-dns" Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.787257 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f68b346-4221-4e2f-bb64-c43040aeb3cd" containerName="dnsmasq-dns" Mar 17 09:30:04 crc kubenswrapper[4813]: E0317 09:30:04.787278 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b380124-90e1-48e8-a3f1-fca49c1adf07" containerName="mariadb-account-create-update" Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.787285 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b380124-90e1-48e8-a3f1-fca49c1adf07" containerName="mariadb-account-create-update" Mar 17 09:30:04 crc kubenswrapper[4813]: E0317 09:30:04.787301 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f68b346-4221-4e2f-bb64-c43040aeb3cd" containerName="init" Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.787307 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f68b346-4221-4e2f-bb64-c43040aeb3cd" containerName="init" Mar 17 09:30:04 crc kubenswrapper[4813]: E0317 09:30:04.787326 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f7359df-58d8-4c25-91b5-d262b088d46e" containerName="mariadb-database-create" Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.787332 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f7359df-58d8-4c25-91b5-d262b088d46e" containerName="mariadb-database-create" Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.787509 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b380124-90e1-48e8-a3f1-fca49c1adf07" containerName="mariadb-account-create-update" Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.787525 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f7359df-58d8-4c25-91b5-d262b088d46e" containerName="mariadb-database-create" Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.787545 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f68b346-4221-4e2f-bb64-c43040aeb3cd" containerName="dnsmasq-dns" Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.790117 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-nrh2h" Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.793240 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bw5js" Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.793666 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.795417 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-nrh2h"] Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.889070 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b8357c2d-7bfc-4296-8d39-7f64f573aa45-db-sync-config-data\") pod \"glance-db-sync-nrh2h\" (UID: \"b8357c2d-7bfc-4296-8d39-7f64f573aa45\") " pod="openstack/glance-db-sync-nrh2h" Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.889124 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8357c2d-7bfc-4296-8d39-7f64f573aa45-config-data\") pod \"glance-db-sync-nrh2h\" (UID: \"b8357c2d-7bfc-4296-8d39-7f64f573aa45\") " pod="openstack/glance-db-sync-nrh2h" Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.889203 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrcd6\" (UniqueName: \"kubernetes.io/projected/b8357c2d-7bfc-4296-8d39-7f64f573aa45-kube-api-access-hrcd6\") pod \"glance-db-sync-nrh2h\" (UID: \"b8357c2d-7bfc-4296-8d39-7f64f573aa45\") " pod="openstack/glance-db-sync-nrh2h" Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.889241 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8357c2d-7bfc-4296-8d39-7f64f573aa45-combined-ca-bundle\") pod \"glance-db-sync-nrh2h\" (UID: \"b8357c2d-7bfc-4296-8d39-7f64f573aa45\") " pod="openstack/glance-db-sync-nrh2h" Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.990570 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8357c2d-7bfc-4296-8d39-7f64f573aa45-combined-ca-bundle\") pod \"glance-db-sync-nrh2h\" (UID: \"b8357c2d-7bfc-4296-8d39-7f64f573aa45\") " pod="openstack/glance-db-sync-nrh2h" Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.990677 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b8357c2d-7bfc-4296-8d39-7f64f573aa45-db-sync-config-data\") pod \"glance-db-sync-nrh2h\" (UID: \"b8357c2d-7bfc-4296-8d39-7f64f573aa45\") " pod="openstack/glance-db-sync-nrh2h" Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.990708 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8357c2d-7bfc-4296-8d39-7f64f573aa45-config-data\") pod \"glance-db-sync-nrh2h\" (UID: \"b8357c2d-7bfc-4296-8d39-7f64f573aa45\") " pod="openstack/glance-db-sync-nrh2h" Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.990774 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrcd6\" (UniqueName: \"kubernetes.io/projected/b8357c2d-7bfc-4296-8d39-7f64f573aa45-kube-api-access-hrcd6\") pod \"glance-db-sync-nrh2h\" (UID: \"b8357c2d-7bfc-4296-8d39-7f64f573aa45\") " pod="openstack/glance-db-sync-nrh2h" Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.998349 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8357c2d-7bfc-4296-8d39-7f64f573aa45-config-data\") pod \"glance-db-sync-nrh2h\" (UID: \"b8357c2d-7bfc-4296-8d39-7f64f573aa45\") " pod="openstack/glance-db-sync-nrh2h" Mar 17 09:30:04 crc kubenswrapper[4813]: I0317 09:30:04.998773 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b8357c2d-7bfc-4296-8d39-7f64f573aa45-db-sync-config-data\") pod \"glance-db-sync-nrh2h\" (UID: \"b8357c2d-7bfc-4296-8d39-7f64f573aa45\") " pod="openstack/glance-db-sync-nrh2h" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.010073 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8357c2d-7bfc-4296-8d39-7f64f573aa45-combined-ca-bundle\") pod \"glance-db-sync-nrh2h\" (UID: \"b8357c2d-7bfc-4296-8d39-7f64f573aa45\") " pod="openstack/glance-db-sync-nrh2h" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.013262 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrcd6\" (UniqueName: \"kubernetes.io/projected/b8357c2d-7bfc-4296-8d39-7f64f573aa45-kube-api-access-hrcd6\") pod \"glance-db-sync-nrh2h\" (UID: \"b8357c2d-7bfc-4296-8d39-7f64f573aa45\") " pod="openstack/glance-db-sync-nrh2h" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.124321 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-nrh2h" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.188351 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-8w5d7" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.195447 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b2d571-5a93-4520-8bab-324e4d27bdc9-operator-scripts\") pod \"d7b2d571-5a93-4520-8bab-324e4d27bdc9\" (UID: \"d7b2d571-5a93-4520-8bab-324e4d27bdc9\") " Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.195589 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgw4x\" (UniqueName: \"kubernetes.io/projected/d7b2d571-5a93-4520-8bab-324e4d27bdc9-kube-api-access-cgw4x\") pod \"d7b2d571-5a93-4520-8bab-324e4d27bdc9\" (UID: \"d7b2d571-5a93-4520-8bab-324e4d27bdc9\") " Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.227867 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d7b2d571-5a93-4520-8bab-324e4d27bdc9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d7b2d571-5a93-4520-8bab-324e4d27bdc9" (UID: "d7b2d571-5a93-4520-8bab-324e4d27bdc9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.242182 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7b2d571-5a93-4520-8bab-324e4d27bdc9-kube-api-access-cgw4x" (OuterVolumeSpecName: "kube-api-access-cgw4x") pod "d7b2d571-5a93-4520-8bab-324e4d27bdc9" (UID: "d7b2d571-5a93-4520-8bab-324e4d27bdc9"). InnerVolumeSpecName "kube-api-access-cgw4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.298547 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d7b2d571-5a93-4520-8bab-324e4d27bdc9-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.298573 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgw4x\" (UniqueName: \"kubernetes.io/projected/d7b2d571-5a93-4520-8bab-324e4d27bdc9-kube-api-access-cgw4x\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.326200 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562330-qjkx6" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.329880 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2d07-account-create-update-4j8f9" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.349415 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-8sdrx" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.374129 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-s6zhg" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.399202 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/47bda8a6-acb6-48df-b15d-1561c26df6cf-secret-volume\") pod \"47bda8a6-acb6-48df-b15d-1561c26df6cf\" (UID: \"47bda8a6-acb6-48df-b15d-1561c26df6cf\") " Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.399513 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vzfbn\" (UniqueName: \"kubernetes.io/projected/47bda8a6-acb6-48df-b15d-1561c26df6cf-kube-api-access-vzfbn\") pod \"47bda8a6-acb6-48df-b15d-1561c26df6cf\" (UID: \"47bda8a6-acb6-48df-b15d-1561c26df6cf\") " Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.399543 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kjwtp\" (UniqueName: \"kubernetes.io/projected/e4b19fd6-39e5-49da-811a-81b0ce1f0c29-kube-api-access-kjwtp\") pod \"e4b19fd6-39e5-49da-811a-81b0ce1f0c29\" (UID: \"e4b19fd6-39e5-49da-811a-81b0ce1f0c29\") " Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.399565 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4296c965-29b4-4676-80ba-959682f0d867-operator-scripts\") pod \"4296c965-29b4-4676-80ba-959682f0d867\" (UID: \"4296c965-29b4-4676-80ba-959682f0d867\") " Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.399586 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4b19fd6-39e5-49da-811a-81b0ce1f0c29-operator-scripts\") pod \"e4b19fd6-39e5-49da-811a-81b0ce1f0c29\" (UID: \"e4b19fd6-39e5-49da-811a-81b0ce1f0c29\") " Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.399629 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqzkj\" (UniqueName: \"kubernetes.io/projected/0cbf9e10-aa11-4c2e-9930-b4f1f451bc65-kube-api-access-nqzkj\") pod \"0cbf9e10-aa11-4c2e-9930-b4f1f451bc65\" (UID: \"0cbf9e10-aa11-4c2e-9930-b4f1f451bc65\") " Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.399664 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7gkq\" (UniqueName: \"kubernetes.io/projected/4296c965-29b4-4676-80ba-959682f0d867-kube-api-access-j7gkq\") pod \"4296c965-29b4-4676-80ba-959682f0d867\" (UID: \"4296c965-29b4-4676-80ba-959682f0d867\") " Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.399691 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/47bda8a6-acb6-48df-b15d-1561c26df6cf-config-volume\") pod \"47bda8a6-acb6-48df-b15d-1561c26df6cf\" (UID: \"47bda8a6-acb6-48df-b15d-1561c26df6cf\") " Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.399746 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0cbf9e10-aa11-4c2e-9930-b4f1f451bc65-operator-scripts\") pod \"0cbf9e10-aa11-4c2e-9930-b4f1f451bc65\" (UID: \"0cbf9e10-aa11-4c2e-9930-b4f1f451bc65\") " Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.400759 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0cbf9e10-aa11-4c2e-9930-b4f1f451bc65-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0cbf9e10-aa11-4c2e-9930-b4f1f451bc65" (UID: "0cbf9e10-aa11-4c2e-9930-b4f1f451bc65"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.401957 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4b19fd6-39e5-49da-811a-81b0ce1f0c29-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "e4b19fd6-39e5-49da-811a-81b0ce1f0c29" (UID: "e4b19fd6-39e5-49da-811a-81b0ce1f0c29"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.407704 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/47bda8a6-acb6-48df-b15d-1561c26df6cf-config-volume" (OuterVolumeSpecName: "config-volume") pod "47bda8a6-acb6-48df-b15d-1561c26df6cf" (UID: "47bda8a6-acb6-48df-b15d-1561c26df6cf"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.409727 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4296c965-29b4-4676-80ba-959682f0d867-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4296c965-29b4-4676-80ba-959682f0d867" (UID: "4296c965-29b4-4676-80ba-959682f0d867"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.410314 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/47bda8a6-acb6-48df-b15d-1561c26df6cf-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "47bda8a6-acb6-48df-b15d-1561c26df6cf" (UID: "47bda8a6-acb6-48df-b15d-1561c26df6cf"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.411686 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cbf9e10-aa11-4c2e-9930-b4f1f451bc65-kube-api-access-nqzkj" (OuterVolumeSpecName: "kube-api-access-nqzkj") pod "0cbf9e10-aa11-4c2e-9930-b4f1f451bc65" (UID: "0cbf9e10-aa11-4c2e-9930-b4f1f451bc65"). InnerVolumeSpecName "kube-api-access-nqzkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.413609 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4296c965-29b4-4676-80ba-959682f0d867-kube-api-access-j7gkq" (OuterVolumeSpecName: "kube-api-access-j7gkq") pod "4296c965-29b4-4676-80ba-959682f0d867" (UID: "4296c965-29b4-4676-80ba-959682f0d867"). InnerVolumeSpecName "kube-api-access-j7gkq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.414140 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4b19fd6-39e5-49da-811a-81b0ce1f0c29-kube-api-access-kjwtp" (OuterVolumeSpecName: "kube-api-access-kjwtp") pod "e4b19fd6-39e5-49da-811a-81b0ce1f0c29" (UID: "e4b19fd6-39e5-49da-811a-81b0ce1f0c29"). InnerVolumeSpecName "kube-api-access-kjwtp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.418161 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47bda8a6-acb6-48df-b15d-1561c26df6cf-kube-api-access-vzfbn" (OuterVolumeSpecName: "kube-api-access-vzfbn") pod "47bda8a6-acb6-48df-b15d-1561c26df6cf" (UID: "47bda8a6-acb6-48df-b15d-1561c26df6cf"). InnerVolumeSpecName "kube-api-access-vzfbn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.501608 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0cbf9e10-aa11-4c2e-9930-b4f1f451bc65-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.501638 4813 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/47bda8a6-acb6-48df-b15d-1561c26df6cf-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.501648 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vzfbn\" (UniqueName: \"kubernetes.io/projected/47bda8a6-acb6-48df-b15d-1561c26df6cf-kube-api-access-vzfbn\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.501658 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kjwtp\" (UniqueName: \"kubernetes.io/projected/e4b19fd6-39e5-49da-811a-81b0ce1f0c29-kube-api-access-kjwtp\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.501667 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4296c965-29b4-4676-80ba-959682f0d867-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.501675 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e4b19fd6-39e5-49da-811a-81b0ce1f0c29-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.501683 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqzkj\" (UniqueName: \"kubernetes.io/projected/0cbf9e10-aa11-4c2e-9930-b4f1f451bc65-kube-api-access-nqzkj\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.501691 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7gkq\" (UniqueName: \"kubernetes.io/projected/4296c965-29b4-4676-80ba-959682f0d867-kube-api-access-j7gkq\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.501700 4813 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/47bda8a6-acb6-48df-b15d-1561c26df6cf-config-volume\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.578499 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-nrh2h"] Mar 17 09:30:05 crc kubenswrapper[4813]: W0317 09:30:05.582956 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb8357c2d_7bfc_4296_8d39_7f64f573aa45.slice/crio-191bbb5ade80db7e5da773b1dca123978dc05c21a79f6feafdf45aff4461849b WatchSource:0}: Error finding container 191bbb5ade80db7e5da773b1dca123978dc05c21a79f6feafdf45aff4461849b: Status 404 returned error can't find the container with id 191bbb5ade80db7e5da773b1dca123978dc05c21a79f6feafdf45aff4461849b Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.720967 4813 generic.go:334] "Generic (PLEG): container finished" podID="339d3b99-3aa5-4c3b-99b2-997de272c26c" containerID="7b352d6cb96798d6f939eb7dc8ae0edd91a09eaeac9d1b8058882634055754d6" exitCode=0 Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.721069 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562330-b5zg5" event={"ID":"339d3b99-3aa5-4c3b-99b2-997de272c26c","Type":"ContainerDied","Data":"7b352d6cb96798d6f939eb7dc8ae0edd91a09eaeac9d1b8058882634055754d6"} Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.723991 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-2d07-account-create-update-4j8f9" event={"ID":"e4b19fd6-39e5-49da-811a-81b0ce1f0c29","Type":"ContainerDied","Data":"366073ccc2c12696b643495e39cf29ec904f5f9813f393dc88491c56fd863885"} Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.724152 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="366073ccc2c12696b643495e39cf29ec904f5f9813f393dc88491c56fd863885" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.724004 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-2d07-account-create-update-4j8f9" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.725932 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-s6zhg" event={"ID":"0cbf9e10-aa11-4c2e-9930-b4f1f451bc65","Type":"ContainerDied","Data":"2d536fd359c6cbcdbcefe8f7ee979ac807ebd5b093342748d3639a6cf64818ee"} Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.725970 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-s6zhg" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.725979 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d536fd359c6cbcdbcefe8f7ee979ac807ebd5b093342748d3639a6cf64818ee" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.727391 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-nrh2h" event={"ID":"b8357c2d-7bfc-4296-8d39-7f64f573aa45","Type":"ContainerStarted","Data":"191bbb5ade80db7e5da773b1dca123978dc05c21a79f6feafdf45aff4461849b"} Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.729640 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562330-qjkx6" event={"ID":"47bda8a6-acb6-48df-b15d-1561c26df6cf","Type":"ContainerDied","Data":"dbb9a3412b7ea1fb0d5e822c5499ee8e31f9a4750c4e89eb5b7ae8536b20f402"} Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.729658 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562330-qjkx6" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.729686 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dbb9a3412b7ea1fb0d5e822c5499ee8e31f9a4750c4e89eb5b7ae8536b20f402" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.731635 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-8sdrx" event={"ID":"4296c965-29b4-4676-80ba-959682f0d867","Type":"ContainerDied","Data":"c997f66df63d205fdef0da50a5a4258eae986e0babbf6ddfe058c7c5ccd57c2e"} Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.731662 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c997f66df63d205fdef0da50a5a4258eae986e0babbf6ddfe058c7c5ccd57c2e" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.731712 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-8sdrx" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.741561 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-8w5d7" event={"ID":"d7b2d571-5a93-4520-8bab-324e4d27bdc9","Type":"ContainerDied","Data":"450ac5009ea4d1336b657eda866aadf5d919fd253fd9d4d7d5f3d39a8bae4ceb"} Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.741593 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="450ac5009ea4d1336b657eda866aadf5d919fd253fd9d4d7d5f3d39a8bae4ceb" Mar 17 09:30:05 crc kubenswrapper[4813]: I0317 09:30:05.741621 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-8w5d7" Mar 17 09:30:07 crc kubenswrapper[4813]: I0317 09:30:07.055141 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Mar 17 09:30:07 crc kubenswrapper[4813]: I0317 09:30:07.134870 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562330-b5zg5" Mar 17 09:30:07 crc kubenswrapper[4813]: I0317 09:30:07.231831 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cg7cv\" (UniqueName: \"kubernetes.io/projected/339d3b99-3aa5-4c3b-99b2-997de272c26c-kube-api-access-cg7cv\") pod \"339d3b99-3aa5-4c3b-99b2-997de272c26c\" (UID: \"339d3b99-3aa5-4c3b-99b2-997de272c26c\") " Mar 17 09:30:07 crc kubenswrapper[4813]: I0317 09:30:07.237778 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/339d3b99-3aa5-4c3b-99b2-997de272c26c-kube-api-access-cg7cv" (OuterVolumeSpecName: "kube-api-access-cg7cv") pod "339d3b99-3aa5-4c3b-99b2-997de272c26c" (UID: "339d3b99-3aa5-4c3b-99b2-997de272c26c"). InnerVolumeSpecName "kube-api-access-cg7cv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:07 crc kubenswrapper[4813]: I0317 09:30:07.333705 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cg7cv\" (UniqueName: \"kubernetes.io/projected/339d3b99-3aa5-4c3b-99b2-997de272c26c-kube-api-access-cg7cv\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:07 crc kubenswrapper[4813]: I0317 09:30:07.559995 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-8sdrx"] Mar 17 09:30:07 crc kubenswrapper[4813]: I0317 09:30:07.567917 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-8sdrx"] Mar 17 09:30:07 crc kubenswrapper[4813]: I0317 09:30:07.763214 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562330-b5zg5" event={"ID":"339d3b99-3aa5-4c3b-99b2-997de272c26c","Type":"ContainerDied","Data":"776c488601eb81dbe228ff8b21929c9a559c1cdba9b8e110794278329856ee94"} Mar 17 09:30:07 crc kubenswrapper[4813]: I0317 09:30:07.763269 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="776c488601eb81dbe228ff8b21929c9a559c1cdba9b8e110794278329856ee94" Mar 17 09:30:07 crc kubenswrapper[4813]: I0317 09:30:07.763299 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562330-b5zg5" Mar 17 09:30:08 crc kubenswrapper[4813]: I0317 09:30:08.188153 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562324-zfp5g"] Mar 17 09:30:08 crc kubenswrapper[4813]: I0317 09:30:08.201710 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562324-zfp5g"] Mar 17 09:30:08 crc kubenswrapper[4813]: I0317 09:30:08.374279 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-etc-swift\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:30:08 crc kubenswrapper[4813]: E0317 09:30:08.374440 4813 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Mar 17 09:30:08 crc kubenswrapper[4813]: E0317 09:30:08.374476 4813 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Mar 17 09:30:08 crc kubenswrapper[4813]: E0317 09:30:08.374540 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-etc-swift podName:76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51 nodeName:}" failed. No retries permitted until 2026-03-17 09:30:24.374517837 +0000 UTC m=+1246.475321356 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-etc-swift") pod "swift-storage-0" (UID: "76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51") : configmap "swift-ring-files" not found Mar 17 09:30:08 crc kubenswrapper[4813]: I0317 09:30:08.741147 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4296c965-29b4-4676-80ba-959682f0d867" path="/var/lib/kubelet/pods/4296c965-29b4-4676-80ba-959682f0d867/volumes" Mar 17 09:30:08 crc kubenswrapper[4813]: I0317 09:30:08.741844 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a6501c3-1a60-4c0d-b190-5a8c4335b527" path="/var/lib/kubelet/pods/9a6501c3-1a60-4c0d-b190-5a8c4335b527/volumes" Mar 17 09:30:08 crc kubenswrapper[4813]: I0317 09:30:08.772993 4813 generic.go:334] "Generic (PLEG): container finished" podID="625a6e48-f069-49a9-b11a-342e12dffece" containerID="aa8f414377f9f29c12c6d9c5723ef631c4ea0afda6b0e5363d36678886bdcd01" exitCode=0 Mar 17 09:30:08 crc kubenswrapper[4813]: I0317 09:30:08.773037 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-rrfrc" event={"ID":"625a6e48-f069-49a9-b11a-342e12dffece","Type":"ContainerDied","Data":"aa8f414377f9f29c12c6d9c5723ef631c4ea0afda6b0e5363d36678886bdcd01"} Mar 17 09:30:09 crc kubenswrapper[4813]: I0317 09:30:09.614363 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-vdc6c" podUID="4090993b-51ce-4ce3-a6d6-a1501ab3ba05" containerName="ovn-controller" probeResult="failure" output=< Mar 17 09:30:09 crc kubenswrapper[4813]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 17 09:30:09 crc kubenswrapper[4813]: > Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.217018 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.409611 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/625a6e48-f069-49a9-b11a-342e12dffece-scripts\") pod \"625a6e48-f069-49a9-b11a-342e12dffece\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.409668 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/625a6e48-f069-49a9-b11a-342e12dffece-swiftconf\") pod \"625a6e48-f069-49a9-b11a-342e12dffece\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.409726 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/625a6e48-f069-49a9-b11a-342e12dffece-ring-data-devices\") pod \"625a6e48-f069-49a9-b11a-342e12dffece\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.410815 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/625a6e48-f069-49a9-b11a-342e12dffece-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "625a6e48-f069-49a9-b11a-342e12dffece" (UID: "625a6e48-f069-49a9-b11a-342e12dffece"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.410864 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/625a6e48-f069-49a9-b11a-342e12dffece-combined-ca-bundle\") pod \"625a6e48-f069-49a9-b11a-342e12dffece\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.410941 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8jhf\" (UniqueName: \"kubernetes.io/projected/625a6e48-f069-49a9-b11a-342e12dffece-kube-api-access-r8jhf\") pod \"625a6e48-f069-49a9-b11a-342e12dffece\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.411342 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/625a6e48-f069-49a9-b11a-342e12dffece-etc-swift\") pod \"625a6e48-f069-49a9-b11a-342e12dffece\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.411389 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/625a6e48-f069-49a9-b11a-342e12dffece-dispersionconf\") pod \"625a6e48-f069-49a9-b11a-342e12dffece\" (UID: \"625a6e48-f069-49a9-b11a-342e12dffece\") " Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.412040 4813 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/625a6e48-f069-49a9-b11a-342e12dffece-ring-data-devices\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.412657 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/625a6e48-f069-49a9-b11a-342e12dffece-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "625a6e48-f069-49a9-b11a-342e12dffece" (UID: "625a6e48-f069-49a9-b11a-342e12dffece"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.418706 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/625a6e48-f069-49a9-b11a-342e12dffece-kube-api-access-r8jhf" (OuterVolumeSpecName: "kube-api-access-r8jhf") pod "625a6e48-f069-49a9-b11a-342e12dffece" (UID: "625a6e48-f069-49a9-b11a-342e12dffece"). InnerVolumeSpecName "kube-api-access-r8jhf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.428295 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/625a6e48-f069-49a9-b11a-342e12dffece-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "625a6e48-f069-49a9-b11a-342e12dffece" (UID: "625a6e48-f069-49a9-b11a-342e12dffece"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.439634 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/625a6e48-f069-49a9-b11a-342e12dffece-scripts" (OuterVolumeSpecName: "scripts") pod "625a6e48-f069-49a9-b11a-342e12dffece" (UID: "625a6e48-f069-49a9-b11a-342e12dffece"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.452257 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/625a6e48-f069-49a9-b11a-342e12dffece-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "625a6e48-f069-49a9-b11a-342e12dffece" (UID: "625a6e48-f069-49a9-b11a-342e12dffece"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.468361 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/625a6e48-f069-49a9-b11a-342e12dffece-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "625a6e48-f069-49a9-b11a-342e12dffece" (UID: "625a6e48-f069-49a9-b11a-342e12dffece"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.515022 4813 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/625a6e48-f069-49a9-b11a-342e12dffece-etc-swift\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.515056 4813 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/625a6e48-f069-49a9-b11a-342e12dffece-dispersionconf\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.515069 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/625a6e48-f069-49a9-b11a-342e12dffece-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.515527 4813 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/625a6e48-f069-49a9-b11a-342e12dffece-swiftconf\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.515545 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/625a6e48-f069-49a9-b11a-342e12dffece-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.515679 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8jhf\" (UniqueName: \"kubernetes.io/projected/625a6e48-f069-49a9-b11a-342e12dffece-kube-api-access-r8jhf\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.788852 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-rrfrc" event={"ID":"625a6e48-f069-49a9-b11a-342e12dffece","Type":"ContainerDied","Data":"f543e5c178e08b935eaf33bcc1d7ea23c030cfb716926fe5ca79788df814c77c"} Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.789084 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f543e5c178e08b935eaf33bcc1d7ea23c030cfb716926fe5ca79788df814c77c" Mar 17 09:30:10 crc kubenswrapper[4813]: I0317 09:30:10.788905 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-rrfrc" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.068624 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-p2b2m"] Mar 17 09:30:11 crc kubenswrapper[4813]: E0317 09:30:11.068959 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="339d3b99-3aa5-4c3b-99b2-997de272c26c" containerName="oc" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.068970 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="339d3b99-3aa5-4c3b-99b2-997de272c26c" containerName="oc" Mar 17 09:30:11 crc kubenswrapper[4813]: E0317 09:30:11.068986 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4296c965-29b4-4676-80ba-959682f0d867" containerName="mariadb-account-create-update" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.068992 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4296c965-29b4-4676-80ba-959682f0d867" containerName="mariadb-account-create-update" Mar 17 09:30:11 crc kubenswrapper[4813]: E0317 09:30:11.069002 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7b2d571-5a93-4520-8bab-324e4d27bdc9" containerName="mariadb-database-create" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.069008 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7b2d571-5a93-4520-8bab-324e4d27bdc9" containerName="mariadb-database-create" Mar 17 09:30:11 crc kubenswrapper[4813]: E0317 09:30:11.069017 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="625a6e48-f069-49a9-b11a-342e12dffece" containerName="swift-ring-rebalance" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.069022 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="625a6e48-f069-49a9-b11a-342e12dffece" containerName="swift-ring-rebalance" Mar 17 09:30:11 crc kubenswrapper[4813]: E0317 09:30:11.069035 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4b19fd6-39e5-49da-811a-81b0ce1f0c29" containerName="mariadb-account-create-update" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.069040 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4b19fd6-39e5-49da-811a-81b0ce1f0c29" containerName="mariadb-account-create-update" Mar 17 09:30:11 crc kubenswrapper[4813]: E0317 09:30:11.069054 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47bda8a6-acb6-48df-b15d-1561c26df6cf" containerName="collect-profiles" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.069059 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="47bda8a6-acb6-48df-b15d-1561c26df6cf" containerName="collect-profiles" Mar 17 09:30:11 crc kubenswrapper[4813]: E0317 09:30:11.069078 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cbf9e10-aa11-4c2e-9930-b4f1f451bc65" containerName="mariadb-database-create" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.069084 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cbf9e10-aa11-4c2e-9930-b4f1f451bc65" containerName="mariadb-database-create" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.069222 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="339d3b99-3aa5-4c3b-99b2-997de272c26c" containerName="oc" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.069235 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7b2d571-5a93-4520-8bab-324e4d27bdc9" containerName="mariadb-database-create" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.069246 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cbf9e10-aa11-4c2e-9930-b4f1f451bc65" containerName="mariadb-database-create" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.069253 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="4296c965-29b4-4676-80ba-959682f0d867" containerName="mariadb-account-create-update" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.069265 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4b19fd6-39e5-49da-811a-81b0ce1f0c29" containerName="mariadb-account-create-update" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.069273 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="625a6e48-f069-49a9-b11a-342e12dffece" containerName="swift-ring-rebalance" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.069282 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="47bda8a6-acb6-48df-b15d-1561c26df6cf" containerName="collect-profiles" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.069834 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-p2b2m" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.073168 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-mariadb-root-db-secret" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.079630 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-p2b2m"] Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.128378 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpxk7\" (UniqueName: \"kubernetes.io/projected/44d9391c-57ca-4760-a02a-e91974db206f-kube-api-access-qpxk7\") pod \"root-account-create-update-p2b2m\" (UID: \"44d9391c-57ca-4760-a02a-e91974db206f\") " pod="openstack/root-account-create-update-p2b2m" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.128487 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44d9391c-57ca-4760-a02a-e91974db206f-operator-scripts\") pod \"root-account-create-update-p2b2m\" (UID: \"44d9391c-57ca-4760-a02a-e91974db206f\") " pod="openstack/root-account-create-update-p2b2m" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.230578 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44d9391c-57ca-4760-a02a-e91974db206f-operator-scripts\") pod \"root-account-create-update-p2b2m\" (UID: \"44d9391c-57ca-4760-a02a-e91974db206f\") " pod="openstack/root-account-create-update-p2b2m" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.230748 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpxk7\" (UniqueName: \"kubernetes.io/projected/44d9391c-57ca-4760-a02a-e91974db206f-kube-api-access-qpxk7\") pod \"root-account-create-update-p2b2m\" (UID: \"44d9391c-57ca-4760-a02a-e91974db206f\") " pod="openstack/root-account-create-update-p2b2m" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.231516 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44d9391c-57ca-4760-a02a-e91974db206f-operator-scripts\") pod \"root-account-create-update-p2b2m\" (UID: \"44d9391c-57ca-4760-a02a-e91974db206f\") " pod="openstack/root-account-create-update-p2b2m" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.259320 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpxk7\" (UniqueName: \"kubernetes.io/projected/44d9391c-57ca-4760-a02a-e91974db206f-kube-api-access-qpxk7\") pod \"root-account-create-update-p2b2m\" (UID: \"44d9391c-57ca-4760-a02a-e91974db206f\") " pod="openstack/root-account-create-update-p2b2m" Mar 17 09:30:11 crc kubenswrapper[4813]: I0317 09:30:11.392402 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-p2b2m" Mar 17 09:30:12 crc kubenswrapper[4813]: I0317 09:30:12.810333 4813 generic.go:334] "Generic (PLEG): container finished" podID="d84c4ab5-7f54-48dd-b45c-03c9dd1ee753" containerID="92494a89ef4e16780b97fced28f929444d04138f37ecab69d9f1fcef95b6682f" exitCode=0 Mar 17 09:30:12 crc kubenswrapper[4813]: I0317 09:30:12.810407 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753","Type":"ContainerDied","Data":"92494a89ef4e16780b97fced28f929444d04138f37ecab69d9f1fcef95b6682f"} Mar 17 09:30:12 crc kubenswrapper[4813]: I0317 09:30:12.814221 4813 generic.go:334] "Generic (PLEG): container finished" podID="2517b76a-309a-4c9e-88e7-6061d528acc8" containerID="d5713bfe8ab956c6c4a2dacaf5fc3b1e6cb4739a7e1e69e2bbd95c16f91dc7df" exitCode=0 Mar 17 09:30:12 crc kubenswrapper[4813]: I0317 09:30:12.814246 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2517b76a-309a-4c9e-88e7-6061d528acc8","Type":"ContainerDied","Data":"d5713bfe8ab956c6c4a2dacaf5fc3b1e6cb4739a7e1e69e2bbd95c16f91dc7df"} Mar 17 09:30:14 crc kubenswrapper[4813]: I0317 09:30:14.113450 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:30:14 crc kubenswrapper[4813]: I0317 09:30:14.113811 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:30:14 crc kubenswrapper[4813]: I0317 09:30:14.113864 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:30:14 crc kubenswrapper[4813]: I0317 09:30:14.114682 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d74d47d3497863468b9c4543e558eb883310048990c8823420967f1910298e01"} pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 09:30:14 crc kubenswrapper[4813]: I0317 09:30:14.114731 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" containerID="cri-o://d74d47d3497863468b9c4543e558eb883310048990c8823420967f1910298e01" gracePeriod=600 Mar 17 09:30:14 crc kubenswrapper[4813]: I0317 09:30:14.629012 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-vdc6c" podUID="4090993b-51ce-4ce3-a6d6-a1501ab3ba05" containerName="ovn-controller" probeResult="failure" output=< Mar 17 09:30:14 crc kubenswrapper[4813]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Mar 17 09:30:14 crc kubenswrapper[4813]: > Mar 17 09:30:14 crc kubenswrapper[4813]: I0317 09:30:14.688570 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:30:14 crc kubenswrapper[4813]: I0317 09:30:14.712922 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-bt47r" Mar 17 09:30:14 crc kubenswrapper[4813]: I0317 09:30:14.939342 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vdc6c-config-xz785"] Mar 17 09:30:14 crc kubenswrapper[4813]: I0317 09:30:14.940309 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vdc6c-config-xz785"] Mar 17 09:30:14 crc kubenswrapper[4813]: I0317 09:30:14.940390 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:14 crc kubenswrapper[4813]: I0317 09:30:14.943256 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.028125 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/eff80576-ea14-4192-a9a0-bebb42e85116-additional-scripts\") pod \"ovn-controller-vdc6c-config-xz785\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.028205 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eff80576-ea14-4192-a9a0-bebb42e85116-scripts\") pod \"ovn-controller-vdc6c-config-xz785\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.028236 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eff80576-ea14-4192-a9a0-bebb42e85116-var-run-ovn\") pod \"ovn-controller-vdc6c-config-xz785\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.028279 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7jxr\" (UniqueName: \"kubernetes.io/projected/eff80576-ea14-4192-a9a0-bebb42e85116-kube-api-access-l7jxr\") pod \"ovn-controller-vdc6c-config-xz785\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.028374 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eff80576-ea14-4192-a9a0-bebb42e85116-var-log-ovn\") pod \"ovn-controller-vdc6c-config-xz785\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.028439 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eff80576-ea14-4192-a9a0-bebb42e85116-var-run\") pod \"ovn-controller-vdc6c-config-xz785\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.130413 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7jxr\" (UniqueName: \"kubernetes.io/projected/eff80576-ea14-4192-a9a0-bebb42e85116-kube-api-access-l7jxr\") pod \"ovn-controller-vdc6c-config-xz785\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.130619 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eff80576-ea14-4192-a9a0-bebb42e85116-var-log-ovn\") pod \"ovn-controller-vdc6c-config-xz785\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.130661 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eff80576-ea14-4192-a9a0-bebb42e85116-var-run\") pod \"ovn-controller-vdc6c-config-xz785\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.130712 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/eff80576-ea14-4192-a9a0-bebb42e85116-additional-scripts\") pod \"ovn-controller-vdc6c-config-xz785\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.130742 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eff80576-ea14-4192-a9a0-bebb42e85116-scripts\") pod \"ovn-controller-vdc6c-config-xz785\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.130760 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eff80576-ea14-4192-a9a0-bebb42e85116-var-run-ovn\") pod \"ovn-controller-vdc6c-config-xz785\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.131030 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eff80576-ea14-4192-a9a0-bebb42e85116-var-run\") pod \"ovn-controller-vdc6c-config-xz785\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.131075 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eff80576-ea14-4192-a9a0-bebb42e85116-var-run-ovn\") pod \"ovn-controller-vdc6c-config-xz785\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.132133 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/eff80576-ea14-4192-a9a0-bebb42e85116-additional-scripts\") pod \"ovn-controller-vdc6c-config-xz785\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.132239 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eff80576-ea14-4192-a9a0-bebb42e85116-var-log-ovn\") pod \"ovn-controller-vdc6c-config-xz785\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.132565 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eff80576-ea14-4192-a9a0-bebb42e85116-scripts\") pod \"ovn-controller-vdc6c-config-xz785\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.161200 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7jxr\" (UniqueName: \"kubernetes.io/projected/eff80576-ea14-4192-a9a0-bebb42e85116-kube-api-access-l7jxr\") pod \"ovn-controller-vdc6c-config-xz785\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.277899 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.848884 4813 generic.go:334] "Generic (PLEG): container finished" podID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerID="d74d47d3497863468b9c4543e558eb883310048990c8823420967f1910298e01" exitCode=0 Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.849380 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerDied","Data":"d74d47d3497863468b9c4543e558eb883310048990c8823420967f1910298e01"} Mar 17 09:30:15 crc kubenswrapper[4813]: I0317 09:30:15.849472 4813 scope.go:117] "RemoveContainer" containerID="c178911cb0823cd7d042791f6122482441f05cc74687451c72a3d02c4974d8e8" Mar 17 09:30:17 crc kubenswrapper[4813]: I0317 09:30:17.289031 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-p2b2m"] Mar 17 09:30:17 crc kubenswrapper[4813]: W0317 09:30:17.296787 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod44d9391c_57ca_4760_a02a_e91974db206f.slice/crio-8f14f54da242108c5c253e29fefdaffcd7ad5dbc375e16c450b3c42b242ac5eb WatchSource:0}: Error finding container 8f14f54da242108c5c253e29fefdaffcd7ad5dbc375e16c450b3c42b242ac5eb: Status 404 returned error can't find the container with id 8f14f54da242108c5c253e29fefdaffcd7ad5dbc375e16c450b3c42b242ac5eb Mar 17 09:30:17 crc kubenswrapper[4813]: I0317 09:30:17.376307 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vdc6c-config-xz785"] Mar 17 09:30:17 crc kubenswrapper[4813]: I0317 09:30:17.867485 4813 generic.go:334] "Generic (PLEG): container finished" podID="44d9391c-57ca-4760-a02a-e91974db206f" containerID="69873978b52fa34c4f053a1af3b3d6f826706713d0fea838103a273204174cbe" exitCode=0 Mar 17 09:30:17 crc kubenswrapper[4813]: I0317 09:30:17.867648 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-p2b2m" event={"ID":"44d9391c-57ca-4760-a02a-e91974db206f","Type":"ContainerDied","Data":"69873978b52fa34c4f053a1af3b3d6f826706713d0fea838103a273204174cbe"} Mar 17 09:30:17 crc kubenswrapper[4813]: I0317 09:30:17.867886 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-p2b2m" event={"ID":"44d9391c-57ca-4760-a02a-e91974db206f","Type":"ContainerStarted","Data":"8f14f54da242108c5c253e29fefdaffcd7ad5dbc375e16c450b3c42b242ac5eb"} Mar 17 09:30:17 crc kubenswrapper[4813]: I0317 09:30:17.870193 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753","Type":"ContainerStarted","Data":"202647fbd1a9ea05b2b013ade8131b58b3d231ff0d159880c4bbdc45f0385dc8"} Mar 17 09:30:17 crc kubenswrapper[4813]: I0317 09:30:17.870459 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:30:17 crc kubenswrapper[4813]: I0317 09:30:17.871744 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vdc6c-config-xz785" event={"ID":"eff80576-ea14-4192-a9a0-bebb42e85116","Type":"ContainerStarted","Data":"415962e7fc7623627abfba1283dd4a72be62912b61ce60043443849f31a9d911"} Mar 17 09:30:17 crc kubenswrapper[4813]: I0317 09:30:17.871786 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vdc6c-config-xz785" event={"ID":"eff80576-ea14-4192-a9a0-bebb42e85116","Type":"ContainerStarted","Data":"48b6306e8b28c194852726730969a8aecbd0df03bdba3168f408bf1c6b902639"} Mar 17 09:30:17 crc kubenswrapper[4813]: I0317 09:30:17.875797 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerStarted","Data":"4ba33e308a10bfb001c8289b81c4de34337bf84d52f75bd8b871b725edd0f79b"} Mar 17 09:30:17 crc kubenswrapper[4813]: I0317 09:30:17.878928 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2517b76a-309a-4c9e-88e7-6061d528acc8","Type":"ContainerStarted","Data":"38f5e639d16fd0e9a3c8edcf8dce798b0c07cc4284524957e965004e5a4ee557"} Mar 17 09:30:17 crc kubenswrapper[4813]: I0317 09:30:17.879691 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 17 09:30:17 crc kubenswrapper[4813]: I0317 09:30:17.882174 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-nrh2h" event={"ID":"b8357c2d-7bfc-4296-8d39-7f64f573aa45","Type":"ContainerStarted","Data":"8d6b382a460b8fffdc35f8457f0856fa2144685a719d1c467f61d7e75a46919f"} Mar 17 09:30:17 crc kubenswrapper[4813]: I0317 09:30:17.919364 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-nrh2h" podStartSLOduration=2.6242176390000003 podStartE2EDuration="13.919340599s" podCreationTimestamp="2026-03-17 09:30:04 +0000 UTC" firstStartedPulling="2026-03-17 09:30:05.585086038 +0000 UTC m=+1227.685889537" lastFinishedPulling="2026-03-17 09:30:16.880208988 +0000 UTC m=+1238.981012497" observedRunningTime="2026-03-17 09:30:17.911766341 +0000 UTC m=+1240.012569870" watchObservedRunningTime="2026-03-17 09:30:17.919340599 +0000 UTC m=+1240.020144098" Mar 17 09:30:17 crc kubenswrapper[4813]: I0317 09:30:17.946001 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=54.083225234 podStartE2EDuration="1m3.945980986s" podCreationTimestamp="2026-03-17 09:29:14 +0000 UTC" firstStartedPulling="2026-03-17 09:29:28.344570165 +0000 UTC m=+1190.445373664" lastFinishedPulling="2026-03-17 09:29:38.207325917 +0000 UTC m=+1200.308129416" observedRunningTime="2026-03-17 09:30:17.94007567 +0000 UTC m=+1240.040879189" watchObservedRunningTime="2026-03-17 09:30:17.945980986 +0000 UTC m=+1240.046784485" Mar 17 09:30:17 crc kubenswrapper[4813]: I0317 09:30:17.959220 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=53.974012811 podStartE2EDuration="1m3.959203853s" podCreationTimestamp="2026-03-17 09:29:14 +0000 UTC" firstStartedPulling="2026-03-17 09:29:27.823103054 +0000 UTC m=+1189.923906553" lastFinishedPulling="2026-03-17 09:29:37.808294096 +0000 UTC m=+1199.909097595" observedRunningTime="2026-03-17 09:30:17.959103509 +0000 UTC m=+1240.059907018" watchObservedRunningTime="2026-03-17 09:30:17.959203853 +0000 UTC m=+1240.060007352" Mar 17 09:30:17 crc kubenswrapper[4813]: I0317 09:30:17.992869 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-vdc6c-config-xz785" podStartSLOduration=3.992854391 podStartE2EDuration="3.992854391s" podCreationTimestamp="2026-03-17 09:30:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:30:17.979246323 +0000 UTC m=+1240.080049822" watchObservedRunningTime="2026-03-17 09:30:17.992854391 +0000 UTC m=+1240.093657890" Mar 17 09:30:18 crc kubenswrapper[4813]: I0317 09:30:18.894736 4813 generic.go:334] "Generic (PLEG): container finished" podID="eff80576-ea14-4192-a9a0-bebb42e85116" containerID="415962e7fc7623627abfba1283dd4a72be62912b61ce60043443849f31a9d911" exitCode=0 Mar 17 09:30:18 crc kubenswrapper[4813]: I0317 09:30:18.896656 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vdc6c-config-xz785" event={"ID":"eff80576-ea14-4192-a9a0-bebb42e85116","Type":"ContainerDied","Data":"415962e7fc7623627abfba1283dd4a72be62912b61ce60043443849f31a9d911"} Mar 17 09:30:19 crc kubenswrapper[4813]: I0317 09:30:19.220118 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-p2b2m" Mar 17 09:30:19 crc kubenswrapper[4813]: I0317 09:30:19.315946 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qpxk7\" (UniqueName: \"kubernetes.io/projected/44d9391c-57ca-4760-a02a-e91974db206f-kube-api-access-qpxk7\") pod \"44d9391c-57ca-4760-a02a-e91974db206f\" (UID: \"44d9391c-57ca-4760-a02a-e91974db206f\") " Mar 17 09:30:19 crc kubenswrapper[4813]: I0317 09:30:19.316030 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44d9391c-57ca-4760-a02a-e91974db206f-operator-scripts\") pod \"44d9391c-57ca-4760-a02a-e91974db206f\" (UID: \"44d9391c-57ca-4760-a02a-e91974db206f\") " Mar 17 09:30:19 crc kubenswrapper[4813]: I0317 09:30:19.316662 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44d9391c-57ca-4760-a02a-e91974db206f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "44d9391c-57ca-4760-a02a-e91974db206f" (UID: "44d9391c-57ca-4760-a02a-e91974db206f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:19 crc kubenswrapper[4813]: I0317 09:30:19.321801 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44d9391c-57ca-4760-a02a-e91974db206f-kube-api-access-qpxk7" (OuterVolumeSpecName: "kube-api-access-qpxk7") pod "44d9391c-57ca-4760-a02a-e91974db206f" (UID: "44d9391c-57ca-4760-a02a-e91974db206f"). InnerVolumeSpecName "kube-api-access-qpxk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:19 crc kubenswrapper[4813]: I0317 09:30:19.418523 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qpxk7\" (UniqueName: \"kubernetes.io/projected/44d9391c-57ca-4760-a02a-e91974db206f-kube-api-access-qpxk7\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:19 crc kubenswrapper[4813]: I0317 09:30:19.418570 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/44d9391c-57ca-4760-a02a-e91974db206f-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:19 crc kubenswrapper[4813]: I0317 09:30:19.623576 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-vdc6c" Mar 17 09:30:19 crc kubenswrapper[4813]: I0317 09:30:19.904692 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-p2b2m" event={"ID":"44d9391c-57ca-4760-a02a-e91974db206f","Type":"ContainerDied","Data":"8f14f54da242108c5c253e29fefdaffcd7ad5dbc375e16c450b3c42b242ac5eb"} Mar 17 09:30:19 crc kubenswrapper[4813]: I0317 09:30:19.904737 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-p2b2m" Mar 17 09:30:19 crc kubenswrapper[4813]: I0317 09:30:19.904764 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f14f54da242108c5c253e29fefdaffcd7ad5dbc375e16c450b3c42b242ac5eb" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.237295 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.333650 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7jxr\" (UniqueName: \"kubernetes.io/projected/eff80576-ea14-4192-a9a0-bebb42e85116-kube-api-access-l7jxr\") pod \"eff80576-ea14-4192-a9a0-bebb42e85116\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.333710 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eff80576-ea14-4192-a9a0-bebb42e85116-scripts\") pod \"eff80576-ea14-4192-a9a0-bebb42e85116\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.333762 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eff80576-ea14-4192-a9a0-bebb42e85116-var-run-ovn\") pod \"eff80576-ea14-4192-a9a0-bebb42e85116\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.333792 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/eff80576-ea14-4192-a9a0-bebb42e85116-additional-scripts\") pod \"eff80576-ea14-4192-a9a0-bebb42e85116\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.333841 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eff80576-ea14-4192-a9a0-bebb42e85116-var-run\") pod \"eff80576-ea14-4192-a9a0-bebb42e85116\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.333891 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eff80576-ea14-4192-a9a0-bebb42e85116-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "eff80576-ea14-4192-a9a0-bebb42e85116" (UID: "eff80576-ea14-4192-a9a0-bebb42e85116"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.333929 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eff80576-ea14-4192-a9a0-bebb42e85116-var-log-ovn\") pod \"eff80576-ea14-4192-a9a0-bebb42e85116\" (UID: \"eff80576-ea14-4192-a9a0-bebb42e85116\") " Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.333960 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eff80576-ea14-4192-a9a0-bebb42e85116-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "eff80576-ea14-4192-a9a0-bebb42e85116" (UID: "eff80576-ea14-4192-a9a0-bebb42e85116"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.333978 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/eff80576-ea14-4192-a9a0-bebb42e85116-var-run" (OuterVolumeSpecName: "var-run") pod "eff80576-ea14-4192-a9a0-bebb42e85116" (UID: "eff80576-ea14-4192-a9a0-bebb42e85116"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.334671 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eff80576-ea14-4192-a9a0-bebb42e85116-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "eff80576-ea14-4192-a9a0-bebb42e85116" (UID: "eff80576-ea14-4192-a9a0-bebb42e85116"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.334859 4813 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/eff80576-ea14-4192-a9a0-bebb42e85116-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.334898 4813 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/eff80576-ea14-4192-a9a0-bebb42e85116-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.334921 4813 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/eff80576-ea14-4192-a9a0-bebb42e85116-var-run\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.334939 4813 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/eff80576-ea14-4192-a9a0-bebb42e85116-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.334974 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eff80576-ea14-4192-a9a0-bebb42e85116-scripts" (OuterVolumeSpecName: "scripts") pod "eff80576-ea14-4192-a9a0-bebb42e85116" (UID: "eff80576-ea14-4192-a9a0-bebb42e85116"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.353046 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eff80576-ea14-4192-a9a0-bebb42e85116-kube-api-access-l7jxr" (OuterVolumeSpecName: "kube-api-access-l7jxr") pod "eff80576-ea14-4192-a9a0-bebb42e85116" (UID: "eff80576-ea14-4192-a9a0-bebb42e85116"). InnerVolumeSpecName "kube-api-access-l7jxr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.436535 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7jxr\" (UniqueName: \"kubernetes.io/projected/eff80576-ea14-4192-a9a0-bebb42e85116-kube-api-access-l7jxr\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.436577 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/eff80576-ea14-4192-a9a0-bebb42e85116-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.474207 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-vdc6c-config-xz785"] Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.481682 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-vdc6c-config-xz785"] Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.569052 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vdc6c-config-jrtfm"] Mar 17 09:30:20 crc kubenswrapper[4813]: E0317 09:30:20.569388 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44d9391c-57ca-4760-a02a-e91974db206f" containerName="mariadb-account-create-update" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.569404 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="44d9391c-57ca-4760-a02a-e91974db206f" containerName="mariadb-account-create-update" Mar 17 09:30:20 crc kubenswrapper[4813]: E0317 09:30:20.569420 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eff80576-ea14-4192-a9a0-bebb42e85116" containerName="ovn-config" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.569428 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="eff80576-ea14-4192-a9a0-bebb42e85116" containerName="ovn-config" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.569662 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="eff80576-ea14-4192-a9a0-bebb42e85116" containerName="ovn-config" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.569675 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="44d9391c-57ca-4760-a02a-e91974db206f" containerName="mariadb-account-create-update" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.570166 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.584870 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vdc6c-config-jrtfm"] Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.640149 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-var-log-ovn\") pod \"ovn-controller-vdc6c-config-jrtfm\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.640271 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-var-run\") pod \"ovn-controller-vdc6c-config-jrtfm\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.640339 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-additional-scripts\") pod \"ovn-controller-vdc6c-config-jrtfm\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.640368 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-var-run-ovn\") pod \"ovn-controller-vdc6c-config-jrtfm\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.640401 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-scripts\") pod \"ovn-controller-vdc6c-config-jrtfm\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.640425 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pcqk2\" (UniqueName: \"kubernetes.io/projected/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-kube-api-access-pcqk2\") pod \"ovn-controller-vdc6c-config-jrtfm\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.742100 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eff80576-ea14-4192-a9a0-bebb42e85116" path="/var/lib/kubelet/pods/eff80576-ea14-4192-a9a0-bebb42e85116/volumes" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.742529 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-var-log-ovn\") pod \"ovn-controller-vdc6c-config-jrtfm\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.742625 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-var-run\") pod \"ovn-controller-vdc6c-config-jrtfm\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.742682 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-additional-scripts\") pod \"ovn-controller-vdc6c-config-jrtfm\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.742710 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-var-run-ovn\") pod \"ovn-controller-vdc6c-config-jrtfm\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.742741 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-scripts\") pod \"ovn-controller-vdc6c-config-jrtfm\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.742763 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pcqk2\" (UniqueName: \"kubernetes.io/projected/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-kube-api-access-pcqk2\") pod \"ovn-controller-vdc6c-config-jrtfm\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.742867 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-var-log-ovn\") pod \"ovn-controller-vdc6c-config-jrtfm\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.742898 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-var-run\") pod \"ovn-controller-vdc6c-config-jrtfm\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.742916 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-var-run-ovn\") pod \"ovn-controller-vdc6c-config-jrtfm\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.743460 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-additional-scripts\") pod \"ovn-controller-vdc6c-config-jrtfm\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.746198 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-scripts\") pod \"ovn-controller-vdc6c-config-jrtfm\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.760140 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pcqk2\" (UniqueName: \"kubernetes.io/projected/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-kube-api-access-pcqk2\") pod \"ovn-controller-vdc6c-config-jrtfm\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.886987 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.912550 4813 scope.go:117] "RemoveContainer" containerID="415962e7fc7623627abfba1283dd4a72be62912b61ce60043443849f31a9d911" Mar 17 09:30:20 crc kubenswrapper[4813]: I0317 09:30:20.912688 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vdc6c-config-xz785" Mar 17 09:30:21 crc kubenswrapper[4813]: W0317 09:30:21.182151 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf0a27f4d_6a63_48e7_b4c3_79036e8118f6.slice/crio-505ccc0ace28fae8aaa2381cb1bf844150b4a926b900b734d7159199e06fbbd5 WatchSource:0}: Error finding container 505ccc0ace28fae8aaa2381cb1bf844150b4a926b900b734d7159199e06fbbd5: Status 404 returned error can't find the container with id 505ccc0ace28fae8aaa2381cb1bf844150b4a926b900b734d7159199e06fbbd5 Mar 17 09:30:21 crc kubenswrapper[4813]: I0317 09:30:21.199243 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vdc6c-config-jrtfm"] Mar 17 09:30:21 crc kubenswrapper[4813]: I0317 09:30:21.924243 4813 generic.go:334] "Generic (PLEG): container finished" podID="f0a27f4d-6a63-48e7-b4c3-79036e8118f6" containerID="cc5738221e82f457b14cb9dfb0e94faa09e5204159938b7bebabc755f0a806d5" exitCode=0 Mar 17 09:30:21 crc kubenswrapper[4813]: I0317 09:30:21.924379 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vdc6c-config-jrtfm" event={"ID":"f0a27f4d-6a63-48e7-b4c3-79036e8118f6","Type":"ContainerDied","Data":"cc5738221e82f457b14cb9dfb0e94faa09e5204159938b7bebabc755f0a806d5"} Mar 17 09:30:21 crc kubenswrapper[4813]: I0317 09:30:21.924974 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vdc6c-config-jrtfm" event={"ID":"f0a27f4d-6a63-48e7-b4c3-79036e8118f6","Type":"ContainerStarted","Data":"505ccc0ace28fae8aaa2381cb1bf844150b4a926b900b734d7159199e06fbbd5"} Mar 17 09:30:22 crc kubenswrapper[4813]: I0317 09:30:22.592830 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-p2b2m"] Mar 17 09:30:22 crc kubenswrapper[4813]: I0317 09:30:22.599003 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-p2b2m"] Mar 17 09:30:22 crc kubenswrapper[4813]: I0317 09:30:22.740680 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44d9391c-57ca-4760-a02a-e91974db206f" path="/var/lib/kubelet/pods/44d9391c-57ca-4760-a02a-e91974db206f/volumes" Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.318205 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.398225 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcqk2\" (UniqueName: \"kubernetes.io/projected/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-kube-api-access-pcqk2\") pod \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.398408 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-scripts\") pod \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.398495 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-var-run\") pod \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.398540 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-var-run" (OuterVolumeSpecName: "var-run") pod "f0a27f4d-6a63-48e7-b4c3-79036e8118f6" (UID: "f0a27f4d-6a63-48e7-b4c3-79036e8118f6"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.398643 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-additional-scripts\") pod \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.398680 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-var-run-ovn\") pod \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.398715 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-var-log-ovn\") pod \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\" (UID: \"f0a27f4d-6a63-48e7-b4c3-79036e8118f6\") " Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.398771 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "f0a27f4d-6a63-48e7-b4c3-79036e8118f6" (UID: "f0a27f4d-6a63-48e7-b4c3-79036e8118f6"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.398918 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "f0a27f4d-6a63-48e7-b4c3-79036e8118f6" (UID: "f0a27f4d-6a63-48e7-b4c3-79036e8118f6"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.399141 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "f0a27f4d-6a63-48e7-b4c3-79036e8118f6" (UID: "f0a27f4d-6a63-48e7-b4c3-79036e8118f6"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.399257 4813 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-additional-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.399282 4813 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-var-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.399301 4813 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-var-log-ovn\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.399321 4813 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-var-run\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.400491 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-scripts" (OuterVolumeSpecName: "scripts") pod "f0a27f4d-6a63-48e7-b4c3-79036e8118f6" (UID: "f0a27f4d-6a63-48e7-b4c3-79036e8118f6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.409007 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-kube-api-access-pcqk2" (OuterVolumeSpecName: "kube-api-access-pcqk2") pod "f0a27f4d-6a63-48e7-b4c3-79036e8118f6" (UID: "f0a27f4d-6a63-48e7-b4c3-79036e8118f6"). InnerVolumeSpecName "kube-api-access-pcqk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.501512 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcqk2\" (UniqueName: \"kubernetes.io/projected/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-kube-api-access-pcqk2\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.501576 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f0a27f4d-6a63-48e7-b4c3-79036e8118f6-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.944672 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vdc6c-config-jrtfm" Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.944933 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vdc6c-config-jrtfm" event={"ID":"f0a27f4d-6a63-48e7-b4c3-79036e8118f6","Type":"ContainerDied","Data":"505ccc0ace28fae8aaa2381cb1bf844150b4a926b900b734d7159199e06fbbd5"} Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.945268 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="505ccc0ace28fae8aaa2381cb1bf844150b4a926b900b734d7159199e06fbbd5" Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.946455 4813 generic.go:334] "Generic (PLEG): container finished" podID="b8357c2d-7bfc-4296-8d39-7f64f573aa45" containerID="8d6b382a460b8fffdc35f8457f0856fa2144685a719d1c467f61d7e75a46919f" exitCode=0 Mar 17 09:30:23 crc kubenswrapper[4813]: I0317 09:30:23.946499 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-nrh2h" event={"ID":"b8357c2d-7bfc-4296-8d39-7f64f573aa45","Type":"ContainerDied","Data":"8d6b382a460b8fffdc35f8457f0856fa2144685a719d1c467f61d7e75a46919f"} Mar 17 09:30:24 crc kubenswrapper[4813]: I0317 09:30:24.409275 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-vdc6c-config-jrtfm"] Mar 17 09:30:24 crc kubenswrapper[4813]: I0317 09:30:24.418452 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-etc-swift\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:30:24 crc kubenswrapper[4813]: I0317 09:30:24.420169 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-vdc6c-config-jrtfm"] Mar 17 09:30:24 crc kubenswrapper[4813]: I0317 09:30:24.425393 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51-etc-swift\") pod \"swift-storage-0\" (UID: \"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51\") " pod="openstack/swift-storage-0" Mar 17 09:30:24 crc kubenswrapper[4813]: I0317 09:30:24.649454 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Mar 17 09:30:24 crc kubenswrapper[4813]: I0317 09:30:24.744751 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0a27f4d-6a63-48e7-b4c3-79036e8118f6" path="/var/lib/kubelet/pods/f0a27f4d-6a63-48e7-b4c3-79036e8118f6/volumes" Mar 17 09:30:25 crc kubenswrapper[4813]: I0317 09:30:25.202812 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Mar 17 09:30:25 crc kubenswrapper[4813]: I0317 09:30:25.320382 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-nrh2h" Mar 17 09:30:25 crc kubenswrapper[4813]: I0317 09:30:25.439785 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b8357c2d-7bfc-4296-8d39-7f64f573aa45-db-sync-config-data\") pod \"b8357c2d-7bfc-4296-8d39-7f64f573aa45\" (UID: \"b8357c2d-7bfc-4296-8d39-7f64f573aa45\") " Mar 17 09:30:25 crc kubenswrapper[4813]: I0317 09:30:25.439951 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8357c2d-7bfc-4296-8d39-7f64f573aa45-config-data\") pod \"b8357c2d-7bfc-4296-8d39-7f64f573aa45\" (UID: \"b8357c2d-7bfc-4296-8d39-7f64f573aa45\") " Mar 17 09:30:25 crc kubenswrapper[4813]: I0317 09:30:25.440056 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrcd6\" (UniqueName: \"kubernetes.io/projected/b8357c2d-7bfc-4296-8d39-7f64f573aa45-kube-api-access-hrcd6\") pod \"b8357c2d-7bfc-4296-8d39-7f64f573aa45\" (UID: \"b8357c2d-7bfc-4296-8d39-7f64f573aa45\") " Mar 17 09:30:25 crc kubenswrapper[4813]: I0317 09:30:25.440142 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8357c2d-7bfc-4296-8d39-7f64f573aa45-combined-ca-bundle\") pod \"b8357c2d-7bfc-4296-8d39-7f64f573aa45\" (UID: \"b8357c2d-7bfc-4296-8d39-7f64f573aa45\") " Mar 17 09:30:25 crc kubenswrapper[4813]: I0317 09:30:25.444906 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8357c2d-7bfc-4296-8d39-7f64f573aa45-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "b8357c2d-7bfc-4296-8d39-7f64f573aa45" (UID: "b8357c2d-7bfc-4296-8d39-7f64f573aa45"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:25 crc kubenswrapper[4813]: I0317 09:30:25.445700 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8357c2d-7bfc-4296-8d39-7f64f573aa45-kube-api-access-hrcd6" (OuterVolumeSpecName: "kube-api-access-hrcd6") pod "b8357c2d-7bfc-4296-8d39-7f64f573aa45" (UID: "b8357c2d-7bfc-4296-8d39-7f64f573aa45"). InnerVolumeSpecName "kube-api-access-hrcd6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:25 crc kubenswrapper[4813]: I0317 09:30:25.467543 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8357c2d-7bfc-4296-8d39-7f64f573aa45-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b8357c2d-7bfc-4296-8d39-7f64f573aa45" (UID: "b8357c2d-7bfc-4296-8d39-7f64f573aa45"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:25 crc kubenswrapper[4813]: I0317 09:30:25.488271 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8357c2d-7bfc-4296-8d39-7f64f573aa45-config-data" (OuterVolumeSpecName: "config-data") pod "b8357c2d-7bfc-4296-8d39-7f64f573aa45" (UID: "b8357c2d-7bfc-4296-8d39-7f64f573aa45"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:25 crc kubenswrapper[4813]: I0317 09:30:25.542511 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8357c2d-7bfc-4296-8d39-7f64f573aa45-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:25 crc kubenswrapper[4813]: I0317 09:30:25.542554 4813 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/b8357c2d-7bfc-4296-8d39-7f64f573aa45-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:25 crc kubenswrapper[4813]: I0317 09:30:25.542567 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8357c2d-7bfc-4296-8d39-7f64f573aa45-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:25 crc kubenswrapper[4813]: I0317 09:30:25.542579 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrcd6\" (UniqueName: \"kubernetes.io/projected/b8357c2d-7bfc-4296-8d39-7f64f573aa45-kube-api-access-hrcd6\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:25 crc kubenswrapper[4813]: I0317 09:30:25.970302 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-nrh2h" event={"ID":"b8357c2d-7bfc-4296-8d39-7f64f573aa45","Type":"ContainerDied","Data":"191bbb5ade80db7e5da773b1dca123978dc05c21a79f6feafdf45aff4461849b"} Mar 17 09:30:25 crc kubenswrapper[4813]: I0317 09:30:25.970797 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="191bbb5ade80db7e5da773b1dca123978dc05c21a79f6feafdf45aff4461849b" Mar 17 09:30:25 crc kubenswrapper[4813]: I0317 09:30:25.970951 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-nrh2h" Mar 17 09:30:25 crc kubenswrapper[4813]: I0317 09:30:25.973391 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51","Type":"ContainerStarted","Data":"60444ebc97f81585e1651fbf222a7510995bb8b5490e271bc81852863b55afce"} Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.353874 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-hxvlc"] Mar 17 09:30:26 crc kubenswrapper[4813]: E0317 09:30:26.354631 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8357c2d-7bfc-4296-8d39-7f64f573aa45" containerName="glance-db-sync" Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.354652 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8357c2d-7bfc-4296-8d39-7f64f573aa45" containerName="glance-db-sync" Mar 17 09:30:26 crc kubenswrapper[4813]: E0317 09:30:26.354664 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0a27f4d-6a63-48e7-b4c3-79036e8118f6" containerName="ovn-config" Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.354672 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0a27f4d-6a63-48e7-b4c3-79036e8118f6" containerName="ovn-config" Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.354891 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8357c2d-7bfc-4296-8d39-7f64f573aa45" containerName="glance-db-sync" Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.354907 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0a27f4d-6a63-48e7-b4c3-79036e8118f6" containerName="ovn-config" Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.355857 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.391797 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-hxvlc"] Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.456068 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-hxvlc\" (UID: \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\") " pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.456106 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-config\") pod \"dnsmasq-dns-5b946c75cc-hxvlc\" (UID: \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\") " pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.456124 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-hxvlc\" (UID: \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\") " pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.456196 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-hxvlc\" (UID: \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\") " pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.456214 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8bns\" (UniqueName: \"kubernetes.io/projected/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-kube-api-access-k8bns\") pod \"dnsmasq-dns-5b946c75cc-hxvlc\" (UID: \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\") " pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.557928 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-hxvlc\" (UID: \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\") " pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.558302 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8bns\" (UniqueName: \"kubernetes.io/projected/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-kube-api-access-k8bns\") pod \"dnsmasq-dns-5b946c75cc-hxvlc\" (UID: \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\") " pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.558395 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-hxvlc\" (UID: \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\") " pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.558420 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-config\") pod \"dnsmasq-dns-5b946c75cc-hxvlc\" (UID: \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\") " pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.558439 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-hxvlc\" (UID: \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\") " pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.559006 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-ovsdbserver-nb\") pod \"dnsmasq-dns-5b946c75cc-hxvlc\" (UID: \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\") " pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.559509 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-ovsdbserver-sb\") pod \"dnsmasq-dns-5b946c75cc-hxvlc\" (UID: \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\") " pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.559549 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-dns-svc\") pod \"dnsmasq-dns-5b946c75cc-hxvlc\" (UID: \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\") " pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.560474 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-config\") pod \"dnsmasq-dns-5b946c75cc-hxvlc\" (UID: \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\") " pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.574115 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8bns\" (UniqueName: \"kubernetes.io/projected/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-kube-api-access-k8bns\") pod \"dnsmasq-dns-5b946c75cc-hxvlc\" (UID: \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\") " pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" Mar 17 09:30:26 crc kubenswrapper[4813]: I0317 09:30:26.686817 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" Mar 17 09:30:27 crc kubenswrapper[4813]: I0317 09:30:27.000297 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51","Type":"ContainerStarted","Data":"4a38f2af772cb9159d1288aeb5750b9291dccf6872f077ce55d4f5f924e6a1bb"} Mar 17 09:30:27 crc kubenswrapper[4813]: I0317 09:30:27.000645 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51","Type":"ContainerStarted","Data":"27a1eb8f4bfb29999fd20e3b4932f5ac3e76a383f6cfc8d5d282ad756abc4bc1"} Mar 17 09:30:27 crc kubenswrapper[4813]: I0317 09:30:27.167256 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-hxvlc"] Mar 17 09:30:27 crc kubenswrapper[4813]: W0317 09:30:27.180020 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod086f03b9_e8c7_4b2c_90c7_8839a4f2b4ef.slice/crio-a5d48bf61c13ee025a82d28793c27792ebea34ca28a064bdcd658f7598ae68a2 WatchSource:0}: Error finding container a5d48bf61c13ee025a82d28793c27792ebea34ca28a064bdcd658f7598ae68a2: Status 404 returned error can't find the container with id a5d48bf61c13ee025a82d28793c27792ebea34ca28a064bdcd658f7598ae68a2 Mar 17 09:30:27 crc kubenswrapper[4813]: I0317 09:30:27.710829 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/root-account-create-update-drprf"] Mar 17 09:30:27 crc kubenswrapper[4813]: I0317 09:30:27.712690 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-drprf" Mar 17 09:30:27 crc kubenswrapper[4813]: I0317 09:30:27.716060 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-cell1-mariadb-root-db-secret" Mar 17 09:30:27 crc kubenswrapper[4813]: I0317 09:30:27.728112 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-drprf"] Mar 17 09:30:27 crc kubenswrapper[4813]: I0317 09:30:27.785613 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8thvd\" (UniqueName: \"kubernetes.io/projected/075476d6-2221-4309-9e08-3d9d4f4bdfe1-kube-api-access-8thvd\") pod \"root-account-create-update-drprf\" (UID: \"075476d6-2221-4309-9e08-3d9d4f4bdfe1\") " pod="openstack/root-account-create-update-drprf" Mar 17 09:30:27 crc kubenswrapper[4813]: I0317 09:30:27.785773 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/075476d6-2221-4309-9e08-3d9d4f4bdfe1-operator-scripts\") pod \"root-account-create-update-drprf\" (UID: \"075476d6-2221-4309-9e08-3d9d4f4bdfe1\") " pod="openstack/root-account-create-update-drprf" Mar 17 09:30:27 crc kubenswrapper[4813]: I0317 09:30:27.886909 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8thvd\" (UniqueName: \"kubernetes.io/projected/075476d6-2221-4309-9e08-3d9d4f4bdfe1-kube-api-access-8thvd\") pod \"root-account-create-update-drprf\" (UID: \"075476d6-2221-4309-9e08-3d9d4f4bdfe1\") " pod="openstack/root-account-create-update-drprf" Mar 17 09:30:27 crc kubenswrapper[4813]: I0317 09:30:27.887020 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/075476d6-2221-4309-9e08-3d9d4f4bdfe1-operator-scripts\") pod \"root-account-create-update-drprf\" (UID: \"075476d6-2221-4309-9e08-3d9d4f4bdfe1\") " pod="openstack/root-account-create-update-drprf" Mar 17 09:30:27 crc kubenswrapper[4813]: I0317 09:30:27.887811 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/075476d6-2221-4309-9e08-3d9d4f4bdfe1-operator-scripts\") pod \"root-account-create-update-drprf\" (UID: \"075476d6-2221-4309-9e08-3d9d4f4bdfe1\") " pod="openstack/root-account-create-update-drprf" Mar 17 09:30:27 crc kubenswrapper[4813]: I0317 09:30:27.936421 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8thvd\" (UniqueName: \"kubernetes.io/projected/075476d6-2221-4309-9e08-3d9d4f4bdfe1-kube-api-access-8thvd\") pod \"root-account-create-update-drprf\" (UID: \"075476d6-2221-4309-9e08-3d9d4f4bdfe1\") " pod="openstack/root-account-create-update-drprf" Mar 17 09:30:28 crc kubenswrapper[4813]: I0317 09:30:28.010575 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51","Type":"ContainerStarted","Data":"b632a7510de1e1e7dccc3a50bd138f9baa3479703a47ccc91e54ad0eb31a0bd4"} Mar 17 09:30:28 crc kubenswrapper[4813]: I0317 09:30:28.010628 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51","Type":"ContainerStarted","Data":"d45fd25c51a076be052b6e5ec3a65890efa6e68347f279f4b059045d1110889d"} Mar 17 09:30:28 crc kubenswrapper[4813]: I0317 09:30:28.011792 4813 generic.go:334] "Generic (PLEG): container finished" podID="086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef" containerID="ec83d45168374c68b33be0f89d81371f7d18619f171609aff389fc083f211959" exitCode=0 Mar 17 09:30:28 crc kubenswrapper[4813]: I0317 09:30:28.011818 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" event={"ID":"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef","Type":"ContainerDied","Data":"ec83d45168374c68b33be0f89d81371f7d18619f171609aff389fc083f211959"} Mar 17 09:30:28 crc kubenswrapper[4813]: I0317 09:30:28.011832 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" event={"ID":"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef","Type":"ContainerStarted","Data":"a5d48bf61c13ee025a82d28793c27792ebea34ca28a064bdcd658f7598ae68a2"} Mar 17 09:30:28 crc kubenswrapper[4813]: I0317 09:30:28.045135 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-drprf" Mar 17 09:30:28 crc kubenswrapper[4813]: I0317 09:30:28.487873 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/root-account-create-update-drprf"] Mar 17 09:30:28 crc kubenswrapper[4813]: W0317 09:30:28.507552 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod075476d6_2221_4309_9e08_3d9d4f4bdfe1.slice/crio-7dd1e215f5074cf174beb90554143e77b041e0b290693d9b142dff39be5826c7 WatchSource:0}: Error finding container 7dd1e215f5074cf174beb90554143e77b041e0b290693d9b142dff39be5826c7: Status 404 returned error can't find the container with id 7dd1e215f5074cf174beb90554143e77b041e0b290693d9b142dff39be5826c7 Mar 17 09:30:29 crc kubenswrapper[4813]: I0317 09:30:29.023258 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51","Type":"ContainerStarted","Data":"4c8ce3087751b1484711bfa3fce17191f2f2e103937aebeb4e52a8984c0bddbf"} Mar 17 09:30:29 crc kubenswrapper[4813]: I0317 09:30:29.023318 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51","Type":"ContainerStarted","Data":"a6c4680e259f0283349d66eb0c454555b27cbf8f1763567d331bb28cc540729a"} Mar 17 09:30:29 crc kubenswrapper[4813]: I0317 09:30:29.025107 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" event={"ID":"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef","Type":"ContainerStarted","Data":"7e2d3acf9a15a11f09b77504bce8d54ec4892e73fd75b245c516def2d2216e4b"} Mar 17 09:30:29 crc kubenswrapper[4813]: I0317 09:30:29.025255 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" Mar 17 09:30:29 crc kubenswrapper[4813]: I0317 09:30:29.027106 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-drprf" event={"ID":"075476d6-2221-4309-9e08-3d9d4f4bdfe1","Type":"ContainerStarted","Data":"19feba5cde4b9c9e1abb5c6c7468776a0c4d71ef6ce95e9dc53af16aa0549546"} Mar 17 09:30:29 crc kubenswrapper[4813]: I0317 09:30:29.027131 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-drprf" event={"ID":"075476d6-2221-4309-9e08-3d9d4f4bdfe1","Type":"ContainerStarted","Data":"7dd1e215f5074cf174beb90554143e77b041e0b290693d9b142dff39be5826c7"} Mar 17 09:30:29 crc kubenswrapper[4813]: I0317 09:30:29.056339 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" podStartSLOduration=3.056323296 podStartE2EDuration="3.056323296s" podCreationTimestamp="2026-03-17 09:30:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:30:29.050930146 +0000 UTC m=+1251.151733645" watchObservedRunningTime="2026-03-17 09:30:29.056323296 +0000 UTC m=+1251.157126795" Mar 17 09:30:29 crc kubenswrapper[4813]: I0317 09:30:29.080547 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/root-account-create-update-drprf" podStartSLOduration=2.080514356 podStartE2EDuration="2.080514356s" podCreationTimestamp="2026-03-17 09:30:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:30:29.072935088 +0000 UTC m=+1251.173738587" watchObservedRunningTime="2026-03-17 09:30:29.080514356 +0000 UTC m=+1251.181317855" Mar 17 09:30:29 crc kubenswrapper[4813]: I0317 09:30:29.342071 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 09:30:30 crc kubenswrapper[4813]: I0317 09:30:30.037704 4813 generic.go:334] "Generic (PLEG): container finished" podID="075476d6-2221-4309-9e08-3d9d4f4bdfe1" containerID="19feba5cde4b9c9e1abb5c6c7468776a0c4d71ef6ce95e9dc53af16aa0549546" exitCode=0 Mar 17 09:30:30 crc kubenswrapper[4813]: I0317 09:30:30.037791 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-drprf" event={"ID":"075476d6-2221-4309-9e08-3d9d4f4bdfe1","Type":"ContainerDied","Data":"19feba5cde4b9c9e1abb5c6c7468776a0c4d71ef6ce95e9dc53af16aa0549546"} Mar 17 09:30:30 crc kubenswrapper[4813]: I0317 09:30:30.047312 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51","Type":"ContainerStarted","Data":"99d7722fc860155ab28ad840a01230931c5c10a45792a1c7df70e77e210ce341"} Mar 17 09:30:30 crc kubenswrapper[4813]: I0317 09:30:30.047352 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51","Type":"ContainerStarted","Data":"0aafcb9f0ae8e434729fadd0fa9c6068f1359ae5e2ac52b412adb3163d637174"} Mar 17 09:30:31 crc kubenswrapper[4813]: I0317 09:30:31.065854 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51","Type":"ContainerStarted","Data":"e093a29bb3030f5ad65c67bd4e2ccccb65f1a309023c427294a85ba052bc0110"} Mar 17 09:30:31 crc kubenswrapper[4813]: I0317 09:30:31.066186 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51","Type":"ContainerStarted","Data":"d352f82ba5cd80348a004f809d1de9b4f2ec22a70484c4ea81483ca479ce5e89"} Mar 17 09:30:31 crc kubenswrapper[4813]: I0317 09:30:31.066208 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51","Type":"ContainerStarted","Data":"3b3c7991035db235c695ec29eeb75261e9850ad88280b646ef0ded5cef300aad"} Mar 17 09:30:31 crc kubenswrapper[4813]: I0317 09:30:31.363112 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-drprf" Mar 17 09:30:31 crc kubenswrapper[4813]: I0317 09:30:31.463682 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8thvd\" (UniqueName: \"kubernetes.io/projected/075476d6-2221-4309-9e08-3d9d4f4bdfe1-kube-api-access-8thvd\") pod \"075476d6-2221-4309-9e08-3d9d4f4bdfe1\" (UID: \"075476d6-2221-4309-9e08-3d9d4f4bdfe1\") " Mar 17 09:30:31 crc kubenswrapper[4813]: I0317 09:30:31.464212 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/075476d6-2221-4309-9e08-3d9d4f4bdfe1-operator-scripts\") pod \"075476d6-2221-4309-9e08-3d9d4f4bdfe1\" (UID: \"075476d6-2221-4309-9e08-3d9d4f4bdfe1\") " Mar 17 09:30:31 crc kubenswrapper[4813]: I0317 09:30:31.465058 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/075476d6-2221-4309-9e08-3d9d4f4bdfe1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "075476d6-2221-4309-9e08-3d9d4f4bdfe1" (UID: "075476d6-2221-4309-9e08-3d9d4f4bdfe1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:31 crc kubenswrapper[4813]: I0317 09:30:31.468102 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/075476d6-2221-4309-9e08-3d9d4f4bdfe1-kube-api-access-8thvd" (OuterVolumeSpecName: "kube-api-access-8thvd") pod "075476d6-2221-4309-9e08-3d9d4f4bdfe1" (UID: "075476d6-2221-4309-9e08-3d9d4f4bdfe1"). InnerVolumeSpecName "kube-api-access-8thvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:31 crc kubenswrapper[4813]: I0317 09:30:31.565714 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8thvd\" (UniqueName: \"kubernetes.io/projected/075476d6-2221-4309-9e08-3d9d4f4bdfe1-kube-api-access-8thvd\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:31 crc kubenswrapper[4813]: I0317 09:30:31.565751 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/075476d6-2221-4309-9e08-3d9d4f4bdfe1-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.079388 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51","Type":"ContainerStarted","Data":"bb791f0aceefb6e0809604d6afc4dfc35f8705ee027975d1d3acf3f2b6412a33"} Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.079672 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51","Type":"ContainerStarted","Data":"9c6f38fe37927c1104707f88b3784a4a44f6905b7fb138fe08ede99256e2c976"} Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.079687 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51","Type":"ContainerStarted","Data":"42dc2468dd8eb182bc9bedebbdda20b974ab7ecddf7399e41b9017a49c05a677"} Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.079699 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51","Type":"ContainerStarted","Data":"e2abe04c9da0d6a76623ed794b35a1b6e70764d0d8c84c08baf0ffae43a24ca8"} Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.080904 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/root-account-create-update-drprf" event={"ID":"075476d6-2221-4309-9e08-3d9d4f4bdfe1","Type":"ContainerDied","Data":"7dd1e215f5074cf174beb90554143e77b041e0b290693d9b142dff39be5826c7"} Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.080934 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7dd1e215f5074cf174beb90554143e77b041e0b290693d9b142dff39be5826c7" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.080950 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/root-account-create-update-drprf" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.449185 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=36.210305927 podStartE2EDuration="41.449169613s" podCreationTimestamp="2026-03-17 09:29:51 +0000 UTC" firstStartedPulling="2026-03-17 09:30:25.215725266 +0000 UTC m=+1247.316528775" lastFinishedPulling="2026-03-17 09:30:30.454588952 +0000 UTC m=+1252.555392461" observedRunningTime="2026-03-17 09:30:32.148457795 +0000 UTC m=+1254.249261314" watchObservedRunningTime="2026-03-17 09:30:32.449169613 +0000 UTC m=+1254.549973112" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.454350 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-hxvlc"] Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.454554 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" podUID="086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef" containerName="dnsmasq-dns" containerID="cri-o://7e2d3acf9a15a11f09b77504bce8d54ec4892e73fd75b245c516def2d2216e4b" gracePeriod=10 Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.491386 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-crbwc"] Mar 17 09:30:32 crc kubenswrapper[4813]: E0317 09:30:32.491895 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="075476d6-2221-4309-9e08-3d9d4f4bdfe1" containerName="mariadb-account-create-update" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.491956 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="075476d6-2221-4309-9e08-3d9d4f4bdfe1" containerName="mariadb-account-create-update" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.492187 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="075476d6-2221-4309-9e08-3d9d4f4bdfe1" containerName="mariadb-account-create-update" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.493037 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.497916 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.502634 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-crbwc"] Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.582705 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-crbwc\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.583037 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-config\") pod \"dnsmasq-dns-7ff5475cc9-crbwc\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.583094 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-crbwc\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.583147 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-crbwc\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.583194 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-crbwc\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.583337 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q2p2\" (UniqueName: \"kubernetes.io/projected/549ff45c-ad4e-4456-bf5a-eec6852665b9-kube-api-access-7q2p2\") pod \"dnsmasq-dns-7ff5475cc9-crbwc\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.685548 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-crbwc\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.685750 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-config\") pod \"dnsmasq-dns-7ff5475cc9-crbwc\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.685784 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-crbwc\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.685807 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-crbwc\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.685839 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-crbwc\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.685882 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q2p2\" (UniqueName: \"kubernetes.io/projected/549ff45c-ad4e-4456-bf5a-eec6852665b9-kube-api-access-7q2p2\") pod \"dnsmasq-dns-7ff5475cc9-crbwc\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.687760 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-dns-swift-storage-0\") pod \"dnsmasq-dns-7ff5475cc9-crbwc\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.688538 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-config\") pod \"dnsmasq-dns-7ff5475cc9-crbwc\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.689248 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-ovsdbserver-nb\") pod \"dnsmasq-dns-7ff5475cc9-crbwc\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.690025 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-ovsdbserver-sb\") pod \"dnsmasq-dns-7ff5475cc9-crbwc\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.690703 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-dns-svc\") pod \"dnsmasq-dns-7ff5475cc9-crbwc\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.724478 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q2p2\" (UniqueName: \"kubernetes.io/projected/549ff45c-ad4e-4456-bf5a-eec6852665b9-kube-api-access-7q2p2\") pod \"dnsmasq-dns-7ff5475cc9-crbwc\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.887268 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.911428 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.990076 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k8bns\" (UniqueName: \"kubernetes.io/projected/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-kube-api-access-k8bns\") pod \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\" (UID: \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\") " Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.990182 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-ovsdbserver-nb\") pod \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\" (UID: \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\") " Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.990331 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-config\") pod \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\" (UID: \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\") " Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.990373 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-dns-svc\") pod \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\" (UID: \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\") " Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.990398 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-ovsdbserver-sb\") pod \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\" (UID: \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\") " Mar 17 09:30:32 crc kubenswrapper[4813]: I0317 09:30:32.997824 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-kube-api-access-k8bns" (OuterVolumeSpecName: "kube-api-access-k8bns") pod "086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef" (UID: "086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef"). InnerVolumeSpecName "kube-api-access-k8bns". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.044173 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef" (UID: "086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.063407 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef" (UID: "086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.089833 4813 generic.go:334] "Generic (PLEG): container finished" podID="086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef" containerID="7e2d3acf9a15a11f09b77504bce8d54ec4892e73fd75b245c516def2d2216e4b" exitCode=0 Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.091279 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.091717 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" event={"ID":"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef","Type":"ContainerDied","Data":"7e2d3acf9a15a11f09b77504bce8d54ec4892e73fd75b245c516def2d2216e4b"} Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.091749 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5b946c75cc-hxvlc" event={"ID":"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef","Type":"ContainerDied","Data":"a5d48bf61c13ee025a82d28793c27792ebea34ca28a064bdcd658f7598ae68a2"} Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.091771 4813 scope.go:117] "RemoveContainer" containerID="7e2d3acf9a15a11f09b77504bce8d54ec4892e73fd75b245c516def2d2216e4b" Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.091774 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-config" (OuterVolumeSpecName: "config") pod "086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef" (UID: "086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.092370 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-config\") pod \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\" (UID: \"086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef\") " Mar 17 09:30:33 crc kubenswrapper[4813]: W0317 09:30:33.092576 4813 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef/volumes/kubernetes.io~configmap/config Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.092620 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-config" (OuterVolumeSpecName: "config") pod "086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef" (UID: "086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.092826 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.092848 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.092863 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k8bns\" (UniqueName: \"kubernetes.io/projected/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-kube-api-access-k8bns\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.092876 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.107991 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef" (UID: "086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.111901 4813 scope.go:117] "RemoveContainer" containerID="ec83d45168374c68b33be0f89d81371f7d18619f171609aff389fc083f211959" Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.147288 4813 scope.go:117] "RemoveContainer" containerID="7e2d3acf9a15a11f09b77504bce8d54ec4892e73fd75b245c516def2d2216e4b" Mar 17 09:30:33 crc kubenswrapper[4813]: E0317 09:30:33.148717 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e2d3acf9a15a11f09b77504bce8d54ec4892e73fd75b245c516def2d2216e4b\": container with ID starting with 7e2d3acf9a15a11f09b77504bce8d54ec4892e73fd75b245c516def2d2216e4b not found: ID does not exist" containerID="7e2d3acf9a15a11f09b77504bce8d54ec4892e73fd75b245c516def2d2216e4b" Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.148756 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e2d3acf9a15a11f09b77504bce8d54ec4892e73fd75b245c516def2d2216e4b"} err="failed to get container status \"7e2d3acf9a15a11f09b77504bce8d54ec4892e73fd75b245c516def2d2216e4b\": rpc error: code = NotFound desc = could not find container \"7e2d3acf9a15a11f09b77504bce8d54ec4892e73fd75b245c516def2d2216e4b\": container with ID starting with 7e2d3acf9a15a11f09b77504bce8d54ec4892e73fd75b245c516def2d2216e4b not found: ID does not exist" Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.148782 4813 scope.go:117] "RemoveContainer" containerID="ec83d45168374c68b33be0f89d81371f7d18619f171609aff389fc083f211959" Mar 17 09:30:33 crc kubenswrapper[4813]: E0317 09:30:33.149350 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec83d45168374c68b33be0f89d81371f7d18619f171609aff389fc083f211959\": container with ID starting with ec83d45168374c68b33be0f89d81371f7d18619f171609aff389fc083f211959 not found: ID does not exist" containerID="ec83d45168374c68b33be0f89d81371f7d18619f171609aff389fc083f211959" Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.149384 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec83d45168374c68b33be0f89d81371f7d18619f171609aff389fc083f211959"} err="failed to get container status \"ec83d45168374c68b33be0f89d81371f7d18619f171609aff389fc083f211959\": rpc error: code = NotFound desc = could not find container \"ec83d45168374c68b33be0f89d81371f7d18619f171609aff389fc083f211959\": container with ID starting with ec83d45168374c68b33be0f89d81371f7d18619f171609aff389fc083f211959 not found: ID does not exist" Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.195257 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.410956 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-crbwc"] Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.431696 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-hxvlc"] Mar 17 09:30:33 crc kubenswrapper[4813]: I0317 09:30:33.444260 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5b946c75cc-hxvlc"] Mar 17 09:30:34 crc kubenswrapper[4813]: I0317 09:30:34.101660 4813 generic.go:334] "Generic (PLEG): container finished" podID="549ff45c-ad4e-4456-bf5a-eec6852665b9" containerID="d30002deb5bc9b97a656826cac0644cd9994c0466037948281bd22e53aa0dcd2" exitCode=0 Mar 17 09:30:34 crc kubenswrapper[4813]: I0317 09:30:34.101824 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" event={"ID":"549ff45c-ad4e-4456-bf5a-eec6852665b9","Type":"ContainerDied","Data":"d30002deb5bc9b97a656826cac0644cd9994c0466037948281bd22e53aa0dcd2"} Mar 17 09:30:34 crc kubenswrapper[4813]: I0317 09:30:34.102154 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" event={"ID":"549ff45c-ad4e-4456-bf5a-eec6852665b9","Type":"ContainerStarted","Data":"7d3434ed12c9880d19cace558c9be3761861a81d2d03cc573a1c7fa04ceba72b"} Mar 17 09:30:34 crc kubenswrapper[4813]: I0317 09:30:34.767594 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef" path="/var/lib/kubelet/pods/086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef/volumes" Mar 17 09:30:35 crc kubenswrapper[4813]: I0317 09:30:35.117500 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" event={"ID":"549ff45c-ad4e-4456-bf5a-eec6852665b9","Type":"ContainerStarted","Data":"af39a9ae05bc43d8ae3e27a24eea19bff217feed6796341341ae24991a1c1dc5"} Mar 17 09:30:35 crc kubenswrapper[4813]: I0317 09:30:35.117972 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:35 crc kubenswrapper[4813]: I0317 09:30:35.158107 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" podStartSLOduration=3.15807089 podStartE2EDuration="3.15807089s" podCreationTimestamp="2026-03-17 09:30:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:30:35.146395452 +0000 UTC m=+1257.247199031" watchObservedRunningTime="2026-03-17 09:30:35.15807089 +0000 UTC m=+1257.258874469" Mar 17 09:30:36 crc kubenswrapper[4813]: I0317 09:30:36.048962 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:30:36 crc kubenswrapper[4813]: I0317 09:30:36.367822 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 17 09:30:37 crc kubenswrapper[4813]: I0317 09:30:37.854952 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-b8kql"] Mar 17 09:30:37 crc kubenswrapper[4813]: E0317 09:30:37.855301 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef" containerName="dnsmasq-dns" Mar 17 09:30:37 crc kubenswrapper[4813]: I0317 09:30:37.855317 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef" containerName="dnsmasq-dns" Mar 17 09:30:37 crc kubenswrapper[4813]: E0317 09:30:37.855344 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef" containerName="init" Mar 17 09:30:37 crc kubenswrapper[4813]: I0317 09:30:37.855353 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef" containerName="init" Mar 17 09:30:37 crc kubenswrapper[4813]: I0317 09:30:37.855548 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="086f03b9-e8c7-4b2c-90c7-8839a4f2b4ef" containerName="dnsmasq-dns" Mar 17 09:30:37 crc kubenswrapper[4813]: I0317 09:30:37.856227 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-b8kql" Mar 17 09:30:37 crc kubenswrapper[4813]: I0317 09:30:37.864870 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-b8kql"] Mar 17 09:30:37 crc kubenswrapper[4813]: I0317 09:30:37.889649 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae41df76-1944-4060-81a1-b28c7c839faf-operator-scripts\") pod \"cinder-db-create-b8kql\" (UID: \"ae41df76-1944-4060-81a1-b28c7c839faf\") " pod="openstack/cinder-db-create-b8kql" Mar 17 09:30:37 crc kubenswrapper[4813]: I0317 09:30:37.889765 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkkh6\" (UniqueName: \"kubernetes.io/projected/ae41df76-1944-4060-81a1-b28c7c839faf-kube-api-access-fkkh6\") pod \"cinder-db-create-b8kql\" (UID: \"ae41df76-1944-4060-81a1-b28c7c839faf\") " pod="openstack/cinder-db-create-b8kql" Mar 17 09:30:37 crc kubenswrapper[4813]: I0317 09:30:37.979614 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-e948-account-create-update-prpwx"] Mar 17 09:30:37 crc kubenswrapper[4813]: I0317 09:30:37.982422 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e948-account-create-update-prpwx" Mar 17 09:30:37 crc kubenswrapper[4813]: I0317 09:30:37.994626 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkkh6\" (UniqueName: \"kubernetes.io/projected/ae41df76-1944-4060-81a1-b28c7c839faf-kube-api-access-fkkh6\") pod \"cinder-db-create-b8kql\" (UID: \"ae41df76-1944-4060-81a1-b28c7c839faf\") " pod="openstack/cinder-db-create-b8kql" Mar 17 09:30:37 crc kubenswrapper[4813]: I0317 09:30:37.994737 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae41df76-1944-4060-81a1-b28c7c839faf-operator-scripts\") pod \"cinder-db-create-b8kql\" (UID: \"ae41df76-1944-4060-81a1-b28c7c839faf\") " pod="openstack/cinder-db-create-b8kql" Mar 17 09:30:37 crc kubenswrapper[4813]: I0317 09:30:37.995294 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-e948-account-create-update-prpwx"] Mar 17 09:30:37 crc kubenswrapper[4813]: I0317 09:30:37.995683 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae41df76-1944-4060-81a1-b28c7c839faf-operator-scripts\") pod \"cinder-db-create-b8kql\" (UID: \"ae41df76-1944-4060-81a1-b28c7c839faf\") " pod="openstack/cinder-db-create-b8kql" Mar 17 09:30:37 crc kubenswrapper[4813]: I0317 09:30:37.996934 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.015708 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkkh6\" (UniqueName: \"kubernetes.io/projected/ae41df76-1944-4060-81a1-b28c7c839faf-kube-api-access-fkkh6\") pod \"cinder-db-create-b8kql\" (UID: \"ae41df76-1944-4060-81a1-b28c7c839faf\") " pod="openstack/cinder-db-create-b8kql" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.061712 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-79mzp"] Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.062703 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-79mzp" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.071141 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-79mzp"] Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.096262 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6cb4aebe-be10-4574-8851-993dc42a3759-operator-scripts\") pod \"barbican-db-create-79mzp\" (UID: \"6cb4aebe-be10-4574-8851-993dc42a3759\") " pod="openstack/barbican-db-create-79mzp" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.096334 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2zfv\" (UniqueName: \"kubernetes.io/projected/6cb4aebe-be10-4574-8851-993dc42a3759-kube-api-access-h2zfv\") pod \"barbican-db-create-79mzp\" (UID: \"6cb4aebe-be10-4574-8851-993dc42a3759\") " pod="openstack/barbican-db-create-79mzp" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.096382 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3666f515-3ed8-486a-a775-eea5c4463580-operator-scripts\") pod \"cinder-e948-account-create-update-prpwx\" (UID: \"3666f515-3ed8-486a-a775-eea5c4463580\") " pod="openstack/cinder-e948-account-create-update-prpwx" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.096562 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bptw\" (UniqueName: \"kubernetes.io/projected/3666f515-3ed8-486a-a775-eea5c4463580-kube-api-access-4bptw\") pod \"cinder-e948-account-create-update-prpwx\" (UID: \"3666f515-3ed8-486a-a775-eea5c4463580\") " pod="openstack/cinder-e948-account-create-update-prpwx" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.160165 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-8098-account-create-update-j4khm"] Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.161146 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8098-account-create-update-j4khm" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.166094 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.172422 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8098-account-create-update-j4khm"] Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.177849 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-b8kql" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.197486 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3666f515-3ed8-486a-a775-eea5c4463580-operator-scripts\") pod \"cinder-e948-account-create-update-prpwx\" (UID: \"3666f515-3ed8-486a-a775-eea5c4463580\") " pod="openstack/cinder-e948-account-create-update-prpwx" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.197562 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bptw\" (UniqueName: \"kubernetes.io/projected/3666f515-3ed8-486a-a775-eea5c4463580-kube-api-access-4bptw\") pod \"cinder-e948-account-create-update-prpwx\" (UID: \"3666f515-3ed8-486a-a775-eea5c4463580\") " pod="openstack/cinder-e948-account-create-update-prpwx" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.197629 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a10880b9-b0fb-4510-a51a-c93bbb6aa563-operator-scripts\") pod \"barbican-8098-account-create-update-j4khm\" (UID: \"a10880b9-b0fb-4510-a51a-c93bbb6aa563\") " pod="openstack/barbican-8098-account-create-update-j4khm" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.197667 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6cb4aebe-be10-4574-8851-993dc42a3759-operator-scripts\") pod \"barbican-db-create-79mzp\" (UID: \"6cb4aebe-be10-4574-8851-993dc42a3759\") " pod="openstack/barbican-db-create-79mzp" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.197710 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2zfv\" (UniqueName: \"kubernetes.io/projected/6cb4aebe-be10-4574-8851-993dc42a3759-kube-api-access-h2zfv\") pod \"barbican-db-create-79mzp\" (UID: \"6cb4aebe-be10-4574-8851-993dc42a3759\") " pod="openstack/barbican-db-create-79mzp" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.197729 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g5hf\" (UniqueName: \"kubernetes.io/projected/a10880b9-b0fb-4510-a51a-c93bbb6aa563-kube-api-access-4g5hf\") pod \"barbican-8098-account-create-update-j4khm\" (UID: \"a10880b9-b0fb-4510-a51a-c93bbb6aa563\") " pod="openstack/barbican-8098-account-create-update-j4khm" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.198455 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3666f515-3ed8-486a-a775-eea5c4463580-operator-scripts\") pod \"cinder-e948-account-create-update-prpwx\" (UID: \"3666f515-3ed8-486a-a775-eea5c4463580\") " pod="openstack/cinder-e948-account-create-update-prpwx" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.199173 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6cb4aebe-be10-4574-8851-993dc42a3759-operator-scripts\") pod \"barbican-db-create-79mzp\" (UID: \"6cb4aebe-be10-4574-8851-993dc42a3759\") " pod="openstack/barbican-db-create-79mzp" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.250181 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2zfv\" (UniqueName: \"kubernetes.io/projected/6cb4aebe-be10-4574-8851-993dc42a3759-kube-api-access-h2zfv\") pod \"barbican-db-create-79mzp\" (UID: \"6cb4aebe-be10-4574-8851-993dc42a3759\") " pod="openstack/barbican-db-create-79mzp" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.256118 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bptw\" (UniqueName: \"kubernetes.io/projected/3666f515-3ed8-486a-a775-eea5c4463580-kube-api-access-4bptw\") pod \"cinder-e948-account-create-update-prpwx\" (UID: \"3666f515-3ed8-486a-a775-eea5c4463580\") " pod="openstack/cinder-e948-account-create-update-prpwx" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.301439 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g5hf\" (UniqueName: \"kubernetes.io/projected/a10880b9-b0fb-4510-a51a-c93bbb6aa563-kube-api-access-4g5hf\") pod \"barbican-8098-account-create-update-j4khm\" (UID: \"a10880b9-b0fb-4510-a51a-c93bbb6aa563\") " pod="openstack/barbican-8098-account-create-update-j4khm" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.301540 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a10880b9-b0fb-4510-a51a-c93bbb6aa563-operator-scripts\") pod \"barbican-8098-account-create-update-j4khm\" (UID: \"a10880b9-b0fb-4510-a51a-c93bbb6aa563\") " pod="openstack/barbican-8098-account-create-update-j4khm" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.302172 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a10880b9-b0fb-4510-a51a-c93bbb6aa563-operator-scripts\") pod \"barbican-8098-account-create-update-j4khm\" (UID: \"a10880b9-b0fb-4510-a51a-c93bbb6aa563\") " pod="openstack/barbican-8098-account-create-update-j4khm" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.306104 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-xlbs6"] Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.307327 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xlbs6" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.313893 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.313940 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e948-account-create-update-prpwx" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.314088 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-tnvpg" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.314199 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.314352 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.327213 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xlbs6"] Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.341347 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g5hf\" (UniqueName: \"kubernetes.io/projected/a10880b9-b0fb-4510-a51a-c93bbb6aa563-kube-api-access-4g5hf\") pod \"barbican-8098-account-create-update-j4khm\" (UID: \"a10880b9-b0fb-4510-a51a-c93bbb6aa563\") " pod="openstack/barbican-8098-account-create-update-j4khm" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.379074 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-79mzp" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.402486 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f531bb05-ff10-4a9b-a8f4-58dc28a8007a-config-data\") pod \"keystone-db-sync-xlbs6\" (UID: \"f531bb05-ff10-4a9b-a8f4-58dc28a8007a\") " pod="openstack/keystone-db-sync-xlbs6" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.402752 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f531bb05-ff10-4a9b-a8f4-58dc28a8007a-combined-ca-bundle\") pod \"keystone-db-sync-xlbs6\" (UID: \"f531bb05-ff10-4a9b-a8f4-58dc28a8007a\") " pod="openstack/keystone-db-sync-xlbs6" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.402781 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thlpr\" (UniqueName: \"kubernetes.io/projected/f531bb05-ff10-4a9b-a8f4-58dc28a8007a-kube-api-access-thlpr\") pod \"keystone-db-sync-xlbs6\" (UID: \"f531bb05-ff10-4a9b-a8f4-58dc28a8007a\") " pod="openstack/keystone-db-sync-xlbs6" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.415423 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-f43a-account-create-update-mxdk5"] Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.416459 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f43a-account-create-update-mxdk5" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.423618 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.430829 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-pn9ft"] Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.431824 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pn9ft" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.448393 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f43a-account-create-update-mxdk5"] Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.476656 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-pn9ft"] Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.500153 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8098-account-create-update-j4khm" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.504412 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nm7v4\" (UniqueName: \"kubernetes.io/projected/34bc865e-b877-44b4-a57d-d7f390f60414-kube-api-access-nm7v4\") pod \"neutron-f43a-account-create-update-mxdk5\" (UID: \"34bc865e-b877-44b4-a57d-d7f390f60414\") " pod="openstack/neutron-f43a-account-create-update-mxdk5" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.504451 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34bc865e-b877-44b4-a57d-d7f390f60414-operator-scripts\") pod \"neutron-f43a-account-create-update-mxdk5\" (UID: \"34bc865e-b877-44b4-a57d-d7f390f60414\") " pod="openstack/neutron-f43a-account-create-update-mxdk5" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.504502 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f531bb05-ff10-4a9b-a8f4-58dc28a8007a-config-data\") pod \"keystone-db-sync-xlbs6\" (UID: \"f531bb05-ff10-4a9b-a8f4-58dc28a8007a\") " pod="openstack/keystone-db-sync-xlbs6" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.504531 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a047928-f15b-4c01-a4ef-b1c4b4d45cc9-operator-scripts\") pod \"neutron-db-create-pn9ft\" (UID: \"5a047928-f15b-4c01-a4ef-b1c4b4d45cc9\") " pod="openstack/neutron-db-create-pn9ft" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.504553 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f531bb05-ff10-4a9b-a8f4-58dc28a8007a-combined-ca-bundle\") pod \"keystone-db-sync-xlbs6\" (UID: \"f531bb05-ff10-4a9b-a8f4-58dc28a8007a\") " pod="openstack/keystone-db-sync-xlbs6" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.504575 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thlpr\" (UniqueName: \"kubernetes.io/projected/f531bb05-ff10-4a9b-a8f4-58dc28a8007a-kube-api-access-thlpr\") pod \"keystone-db-sync-xlbs6\" (UID: \"f531bb05-ff10-4a9b-a8f4-58dc28a8007a\") " pod="openstack/keystone-db-sync-xlbs6" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.504590 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7pndq\" (UniqueName: \"kubernetes.io/projected/5a047928-f15b-4c01-a4ef-b1c4b4d45cc9-kube-api-access-7pndq\") pod \"neutron-db-create-pn9ft\" (UID: \"5a047928-f15b-4c01-a4ef-b1c4b4d45cc9\") " pod="openstack/neutron-db-create-pn9ft" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.529538 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f531bb05-ff10-4a9b-a8f4-58dc28a8007a-combined-ca-bundle\") pod \"keystone-db-sync-xlbs6\" (UID: \"f531bb05-ff10-4a9b-a8f4-58dc28a8007a\") " pod="openstack/keystone-db-sync-xlbs6" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.532215 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thlpr\" (UniqueName: \"kubernetes.io/projected/f531bb05-ff10-4a9b-a8f4-58dc28a8007a-kube-api-access-thlpr\") pod \"keystone-db-sync-xlbs6\" (UID: \"f531bb05-ff10-4a9b-a8f4-58dc28a8007a\") " pod="openstack/keystone-db-sync-xlbs6" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.535764 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f531bb05-ff10-4a9b-a8f4-58dc28a8007a-config-data\") pod \"keystone-db-sync-xlbs6\" (UID: \"f531bb05-ff10-4a9b-a8f4-58dc28a8007a\") " pod="openstack/keystone-db-sync-xlbs6" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.605676 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nm7v4\" (UniqueName: \"kubernetes.io/projected/34bc865e-b877-44b4-a57d-d7f390f60414-kube-api-access-nm7v4\") pod \"neutron-f43a-account-create-update-mxdk5\" (UID: \"34bc865e-b877-44b4-a57d-d7f390f60414\") " pod="openstack/neutron-f43a-account-create-update-mxdk5" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.605736 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34bc865e-b877-44b4-a57d-d7f390f60414-operator-scripts\") pod \"neutron-f43a-account-create-update-mxdk5\" (UID: \"34bc865e-b877-44b4-a57d-d7f390f60414\") " pod="openstack/neutron-f43a-account-create-update-mxdk5" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.605813 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a047928-f15b-4c01-a4ef-b1c4b4d45cc9-operator-scripts\") pod \"neutron-db-create-pn9ft\" (UID: \"5a047928-f15b-4c01-a4ef-b1c4b4d45cc9\") " pod="openstack/neutron-db-create-pn9ft" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.605861 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7pndq\" (UniqueName: \"kubernetes.io/projected/5a047928-f15b-4c01-a4ef-b1c4b4d45cc9-kube-api-access-7pndq\") pod \"neutron-db-create-pn9ft\" (UID: \"5a047928-f15b-4c01-a4ef-b1c4b4d45cc9\") " pod="openstack/neutron-db-create-pn9ft" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.606503 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34bc865e-b877-44b4-a57d-d7f390f60414-operator-scripts\") pod \"neutron-f43a-account-create-update-mxdk5\" (UID: \"34bc865e-b877-44b4-a57d-d7f390f60414\") " pod="openstack/neutron-f43a-account-create-update-mxdk5" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.606904 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a047928-f15b-4c01-a4ef-b1c4b4d45cc9-operator-scripts\") pod \"neutron-db-create-pn9ft\" (UID: \"5a047928-f15b-4c01-a4ef-b1c4b4d45cc9\") " pod="openstack/neutron-db-create-pn9ft" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.633192 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nm7v4\" (UniqueName: \"kubernetes.io/projected/34bc865e-b877-44b4-a57d-d7f390f60414-kube-api-access-nm7v4\") pod \"neutron-f43a-account-create-update-mxdk5\" (UID: \"34bc865e-b877-44b4-a57d-d7f390f60414\") " pod="openstack/neutron-f43a-account-create-update-mxdk5" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.636015 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7pndq\" (UniqueName: \"kubernetes.io/projected/5a047928-f15b-4c01-a4ef-b1c4b4d45cc9-kube-api-access-7pndq\") pod \"neutron-db-create-pn9ft\" (UID: \"5a047928-f15b-4c01-a4ef-b1c4b4d45cc9\") " pod="openstack/neutron-db-create-pn9ft" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.651881 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-tnvpg" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.661140 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xlbs6" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.817060 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f43a-account-create-update-mxdk5" Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.836145 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-b8kql"] Mar 17 09:30:38 crc kubenswrapper[4813]: I0317 09:30:38.847180 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pn9ft" Mar 17 09:30:38 crc kubenswrapper[4813]: W0317 09:30:38.854663 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podae41df76_1944_4060_81a1_b28c7c839faf.slice/crio-589ad2a3e9b14dec62a01273af8a7676ac7c79955283508d520f0417fcc951f2 WatchSource:0}: Error finding container 589ad2a3e9b14dec62a01273af8a7676ac7c79955283508d520f0417fcc951f2: Status 404 returned error can't find the container with id 589ad2a3e9b14dec62a01273af8a7676ac7c79955283508d520f0417fcc951f2 Mar 17 09:30:39 crc kubenswrapper[4813]: I0317 09:30:39.024982 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-e948-account-create-update-prpwx"] Mar 17 09:30:39 crc kubenswrapper[4813]: W0317 09:30:39.033493 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6cb4aebe_be10_4574_8851_993dc42a3759.slice/crio-a2393c81f55a914413980d991bfeb3490466d5a43f8dddb2b99cfa493dedb25f WatchSource:0}: Error finding container a2393c81f55a914413980d991bfeb3490466d5a43f8dddb2b99cfa493dedb25f: Status 404 returned error can't find the container with id a2393c81f55a914413980d991bfeb3490466d5a43f8dddb2b99cfa493dedb25f Mar 17 09:30:39 crc kubenswrapper[4813]: I0317 09:30:39.037724 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-79mzp"] Mar 17 09:30:39 crc kubenswrapper[4813]: I0317 09:30:39.043577 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Mar 17 09:30:39 crc kubenswrapper[4813]: I0317 09:30:39.130047 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-8098-account-create-update-j4khm"] Mar 17 09:30:39 crc kubenswrapper[4813]: W0317 09:30:39.133104 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda10880b9_b0fb_4510_a51a_c93bbb6aa563.slice/crio-bdddcd7cb666fe4046e482824738cd03a67418539cb028a9186e53673ce67707 WatchSource:0}: Error finding container bdddcd7cb666fe4046e482824738cd03a67418539cb028a9186e53673ce67707: Status 404 returned error can't find the container with id bdddcd7cb666fe4046e482824738cd03a67418539cb028a9186e53673ce67707 Mar 17 09:30:39 crc kubenswrapper[4813]: I0317 09:30:39.139209 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Mar 17 09:30:39 crc kubenswrapper[4813]: I0317 09:30:39.186210 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e948-account-create-update-prpwx" event={"ID":"3666f515-3ed8-486a-a775-eea5c4463580","Type":"ContainerStarted","Data":"88f24b266ffaba1aa9cf9eb521c0817c0ab693b1bc3770f24f029f5a2287d51b"} Mar 17 09:30:39 crc kubenswrapper[4813]: I0317 09:30:39.195128 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-79mzp" event={"ID":"6cb4aebe-be10-4574-8851-993dc42a3759","Type":"ContainerStarted","Data":"a2393c81f55a914413980d991bfeb3490466d5a43f8dddb2b99cfa493dedb25f"} Mar 17 09:30:39 crc kubenswrapper[4813]: I0317 09:30:39.197059 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-b8kql" event={"ID":"ae41df76-1944-4060-81a1-b28c7c839faf","Type":"ContainerStarted","Data":"589ad2a3e9b14dec62a01273af8a7676ac7c79955283508d520f0417fcc951f2"} Mar 17 09:30:39 crc kubenswrapper[4813]: I0317 09:30:39.199500 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8098-account-create-update-j4khm" event={"ID":"a10880b9-b0fb-4510-a51a-c93bbb6aa563","Type":"ContainerStarted","Data":"bdddcd7cb666fe4046e482824738cd03a67418539cb028a9186e53673ce67707"} Mar 17 09:30:39 crc kubenswrapper[4813]: I0317 09:30:39.251708 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-xlbs6"] Mar 17 09:30:39 crc kubenswrapper[4813]: W0317 09:30:39.256312 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf531bb05_ff10_4a9b_a8f4_58dc28a8007a.slice/crio-8823cfcec655ecd7985551e2040e420b3d2ab43d411c25adeea3d32efa6b967a WatchSource:0}: Error finding container 8823cfcec655ecd7985551e2040e420b3d2ab43d411c25adeea3d32efa6b967a: Status 404 returned error can't find the container with id 8823cfcec655ecd7985551e2040e420b3d2ab43d411c25adeea3d32efa6b967a Mar 17 09:30:39 crc kubenswrapper[4813]: I0317 09:30:39.392878 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-f43a-account-create-update-mxdk5"] Mar 17 09:30:39 crc kubenswrapper[4813]: I0317 09:30:39.399238 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Mar 17 09:30:39 crc kubenswrapper[4813]: W0317 09:30:39.465306 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5a047928_f15b_4c01_a4ef_b1c4b4d45cc9.slice/crio-d353d5e812f81cdcf7741643f7148d3c13bbf5a360ed58cf303df722cdb60c27 WatchSource:0}: Error finding container d353d5e812f81cdcf7741643f7148d3c13bbf5a360ed58cf303df722cdb60c27: Status 404 returned error can't find the container with id d353d5e812f81cdcf7741643f7148d3c13bbf5a360ed58cf303df722cdb60c27 Mar 17 09:30:39 crc kubenswrapper[4813]: I0317 09:30:39.465965 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-pn9ft"] Mar 17 09:30:39 crc kubenswrapper[4813]: I0317 09:30:39.894052 4813 scope.go:117] "RemoveContainer" containerID="e7d340a3db815657bdf95e7b75ba6d139f9ce745fe0d402a1d6530f938945399" Mar 17 09:30:40 crc kubenswrapper[4813]: I0317 09:30:40.216614 4813 generic.go:334] "Generic (PLEG): container finished" podID="ae41df76-1944-4060-81a1-b28c7c839faf" containerID="3759fe17e8df5399c3f9362f787395f71f578aa9dd86d678642fc1e444352438" exitCode=0 Mar 17 09:30:40 crc kubenswrapper[4813]: I0317 09:30:40.216654 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-b8kql" event={"ID":"ae41df76-1944-4060-81a1-b28c7c839faf","Type":"ContainerDied","Data":"3759fe17e8df5399c3f9362f787395f71f578aa9dd86d678642fc1e444352438"} Mar 17 09:30:40 crc kubenswrapper[4813]: I0317 09:30:40.218648 4813 generic.go:334] "Generic (PLEG): container finished" podID="a10880b9-b0fb-4510-a51a-c93bbb6aa563" containerID="c6452ad9a30c1975243728d5c7fab8eae68073e39d7f4fdf834c41edbbf86c44" exitCode=0 Mar 17 09:30:40 crc kubenswrapper[4813]: I0317 09:30:40.218728 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8098-account-create-update-j4khm" event={"ID":"a10880b9-b0fb-4510-a51a-c93bbb6aa563","Type":"ContainerDied","Data":"c6452ad9a30c1975243728d5c7fab8eae68073e39d7f4fdf834c41edbbf86c44"} Mar 17 09:30:40 crc kubenswrapper[4813]: I0317 09:30:40.220321 4813 generic.go:334] "Generic (PLEG): container finished" podID="3666f515-3ed8-486a-a775-eea5c4463580" containerID="87f5d41a069b50be3e09350e28a6e33d90b1d628153073fab9b9017b3d1e78af" exitCode=0 Mar 17 09:30:40 crc kubenswrapper[4813]: I0317 09:30:40.220363 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e948-account-create-update-prpwx" event={"ID":"3666f515-3ed8-486a-a775-eea5c4463580","Type":"ContainerDied","Data":"87f5d41a069b50be3e09350e28a6e33d90b1d628153073fab9b9017b3d1e78af"} Mar 17 09:30:40 crc kubenswrapper[4813]: I0317 09:30:40.222208 4813 generic.go:334] "Generic (PLEG): container finished" podID="34bc865e-b877-44b4-a57d-d7f390f60414" containerID="bed2240cb56ec958df9556f000c568cf41aaf1e047d6713adee45fc9e40288b1" exitCode=0 Mar 17 09:30:40 crc kubenswrapper[4813]: I0317 09:30:40.222293 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f43a-account-create-update-mxdk5" event={"ID":"34bc865e-b877-44b4-a57d-d7f390f60414","Type":"ContainerDied","Data":"bed2240cb56ec958df9556f000c568cf41aaf1e047d6713adee45fc9e40288b1"} Mar 17 09:30:40 crc kubenswrapper[4813]: I0317 09:30:40.222311 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f43a-account-create-update-mxdk5" event={"ID":"34bc865e-b877-44b4-a57d-d7f390f60414","Type":"ContainerStarted","Data":"49df134cffcb0cf867f2189e839d8f0a5b5207e58e0319028fd99c6eb27dfcaa"} Mar 17 09:30:40 crc kubenswrapper[4813]: I0317 09:30:40.223434 4813 generic.go:334] "Generic (PLEG): container finished" podID="6cb4aebe-be10-4574-8851-993dc42a3759" containerID="cced6e2efb9888502f8215f475b217859c218afc2a981b90b782f833493ba942" exitCode=0 Mar 17 09:30:40 crc kubenswrapper[4813]: I0317 09:30:40.223516 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-79mzp" event={"ID":"6cb4aebe-be10-4574-8851-993dc42a3759","Type":"ContainerDied","Data":"cced6e2efb9888502f8215f475b217859c218afc2a981b90b782f833493ba942"} Mar 17 09:30:40 crc kubenswrapper[4813]: I0317 09:30:40.224518 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xlbs6" event={"ID":"f531bb05-ff10-4a9b-a8f4-58dc28a8007a","Type":"ContainerStarted","Data":"8823cfcec655ecd7985551e2040e420b3d2ab43d411c25adeea3d32efa6b967a"} Mar 17 09:30:40 crc kubenswrapper[4813]: I0317 09:30:40.226162 4813 generic.go:334] "Generic (PLEG): container finished" podID="5a047928-f15b-4c01-a4ef-b1c4b4d45cc9" containerID="18dad9114889c2dcbe23adb0769418a921cf69093eb881946fbd29b93b802eb0" exitCode=0 Mar 17 09:30:40 crc kubenswrapper[4813]: I0317 09:30:40.226189 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pn9ft" event={"ID":"5a047928-f15b-4c01-a4ef-b1c4b4d45cc9","Type":"ContainerDied","Data":"18dad9114889c2dcbe23adb0769418a921cf69093eb881946fbd29b93b802eb0"} Mar 17 09:30:40 crc kubenswrapper[4813]: I0317 09:30:40.226202 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pn9ft" event={"ID":"5a047928-f15b-4c01-a4ef-b1c4b4d45cc9","Type":"ContainerStarted","Data":"d353d5e812f81cdcf7741643f7148d3c13bbf5a360ed58cf303df722cdb60c27"} Mar 17 09:30:42 crc kubenswrapper[4813]: I0317 09:30:42.913803 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:30:42 crc kubenswrapper[4813]: I0317 09:30:42.994178 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-b25vw"] Mar 17 09:30:42 crc kubenswrapper[4813]: I0317 09:30:42.994552 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-b25vw" podUID="8552c885-2651-4424-b0d6-f75a5b7487dc" containerName="dnsmasq-dns" containerID="cri-o://0f2903c7e92e7817c2c04a7025fca74674207ad263f5ddacaf07b970622b3116" gracePeriod=10 Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.265439 4813 generic.go:334] "Generic (PLEG): container finished" podID="8552c885-2651-4424-b0d6-f75a5b7487dc" containerID="0f2903c7e92e7817c2c04a7025fca74674207ad263f5ddacaf07b970622b3116" exitCode=0 Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.265541 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-b25vw" event={"ID":"8552c885-2651-4424-b0d6-f75a5b7487dc","Type":"ContainerDied","Data":"0f2903c7e92e7817c2c04a7025fca74674207ad263f5ddacaf07b970622b3116"} Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.802634 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f43a-account-create-update-mxdk5" Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.835347 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-b8kql" Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.849471 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8098-account-create-update-j4khm" Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.865652 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pn9ft" Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.873825 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e948-account-create-update-prpwx" Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.887069 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-79mzp" Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.897549 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fkkh6\" (UniqueName: \"kubernetes.io/projected/ae41df76-1944-4060-81a1-b28c7c839faf-kube-api-access-fkkh6\") pod \"ae41df76-1944-4060-81a1-b28c7c839faf\" (UID: \"ae41df76-1944-4060-81a1-b28c7c839faf\") " Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.897621 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7pndq\" (UniqueName: \"kubernetes.io/projected/5a047928-f15b-4c01-a4ef-b1c4b4d45cc9-kube-api-access-7pndq\") pod \"5a047928-f15b-4c01-a4ef-b1c4b4d45cc9\" (UID: \"5a047928-f15b-4c01-a4ef-b1c4b4d45cc9\") " Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.897710 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a10880b9-b0fb-4510-a51a-c93bbb6aa563-operator-scripts\") pod \"a10880b9-b0fb-4510-a51a-c93bbb6aa563\" (UID: \"a10880b9-b0fb-4510-a51a-c93bbb6aa563\") " Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.897758 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3666f515-3ed8-486a-a775-eea5c4463580-operator-scripts\") pod \"3666f515-3ed8-486a-a775-eea5c4463580\" (UID: \"3666f515-3ed8-486a-a775-eea5c4463580\") " Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.897869 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nm7v4\" (UniqueName: \"kubernetes.io/projected/34bc865e-b877-44b4-a57d-d7f390f60414-kube-api-access-nm7v4\") pod \"34bc865e-b877-44b4-a57d-d7f390f60414\" (UID: \"34bc865e-b877-44b4-a57d-d7f390f60414\") " Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.897938 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34bc865e-b877-44b4-a57d-d7f390f60414-operator-scripts\") pod \"34bc865e-b877-44b4-a57d-d7f390f60414\" (UID: \"34bc865e-b877-44b4-a57d-d7f390f60414\") " Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.897989 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a047928-f15b-4c01-a4ef-b1c4b4d45cc9-operator-scripts\") pod \"5a047928-f15b-4c01-a4ef-b1c4b4d45cc9\" (UID: \"5a047928-f15b-4c01-a4ef-b1c4b4d45cc9\") " Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.898016 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4g5hf\" (UniqueName: \"kubernetes.io/projected/a10880b9-b0fb-4510-a51a-c93bbb6aa563-kube-api-access-4g5hf\") pod \"a10880b9-b0fb-4510-a51a-c93bbb6aa563\" (UID: \"a10880b9-b0fb-4510-a51a-c93bbb6aa563\") " Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.898063 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae41df76-1944-4060-81a1-b28c7c839faf-operator-scripts\") pod \"ae41df76-1944-4060-81a1-b28c7c839faf\" (UID: \"ae41df76-1944-4060-81a1-b28c7c839faf\") " Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.898099 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bptw\" (UniqueName: \"kubernetes.io/projected/3666f515-3ed8-486a-a775-eea5c4463580-kube-api-access-4bptw\") pod \"3666f515-3ed8-486a-a775-eea5c4463580\" (UID: \"3666f515-3ed8-486a-a775-eea5c4463580\") " Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.899393 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a10880b9-b0fb-4510-a51a-c93bbb6aa563-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a10880b9-b0fb-4510-a51a-c93bbb6aa563" (UID: "a10880b9-b0fb-4510-a51a-c93bbb6aa563"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.899908 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5a047928-f15b-4c01-a4ef-b1c4b4d45cc9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5a047928-f15b-4c01-a4ef-b1c4b4d45cc9" (UID: "5a047928-f15b-4c01-a4ef-b1c4b4d45cc9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.900835 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/34bc865e-b877-44b4-a57d-d7f390f60414-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "34bc865e-b877-44b4-a57d-d7f390f60414" (UID: "34bc865e-b877-44b4-a57d-d7f390f60414"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.900957 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-b25vw" Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.904358 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3666f515-3ed8-486a-a775-eea5c4463580-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3666f515-3ed8-486a-a775-eea5c4463580" (UID: "3666f515-3ed8-486a-a775-eea5c4463580"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.905229 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ae41df76-1944-4060-81a1-b28c7c839faf-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ae41df76-1944-4060-81a1-b28c7c839faf" (UID: "ae41df76-1944-4060-81a1-b28c7c839faf"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.908067 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/34bc865e-b877-44b4-a57d-d7f390f60414-kube-api-access-nm7v4" (OuterVolumeSpecName: "kube-api-access-nm7v4") pod "34bc865e-b877-44b4-a57d-d7f390f60414" (UID: "34bc865e-b877-44b4-a57d-d7f390f60414"). InnerVolumeSpecName "kube-api-access-nm7v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.911041 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a047928-f15b-4c01-a4ef-b1c4b4d45cc9-kube-api-access-7pndq" (OuterVolumeSpecName: "kube-api-access-7pndq") pod "5a047928-f15b-4c01-a4ef-b1c4b4d45cc9" (UID: "5a047928-f15b-4c01-a4ef-b1c4b4d45cc9"). InnerVolumeSpecName "kube-api-access-7pndq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.912122 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3666f515-3ed8-486a-a775-eea5c4463580-kube-api-access-4bptw" (OuterVolumeSpecName: "kube-api-access-4bptw") pod "3666f515-3ed8-486a-a775-eea5c4463580" (UID: "3666f515-3ed8-486a-a775-eea5c4463580"). InnerVolumeSpecName "kube-api-access-4bptw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.914671 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae41df76-1944-4060-81a1-b28c7c839faf-kube-api-access-fkkh6" (OuterVolumeSpecName: "kube-api-access-fkkh6") pod "ae41df76-1944-4060-81a1-b28c7c839faf" (UID: "ae41df76-1944-4060-81a1-b28c7c839faf"). InnerVolumeSpecName "kube-api-access-fkkh6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.917220 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a10880b9-b0fb-4510-a51a-c93bbb6aa563-kube-api-access-4g5hf" (OuterVolumeSpecName: "kube-api-access-4g5hf") pod "a10880b9-b0fb-4510-a51a-c93bbb6aa563" (UID: "a10880b9-b0fb-4510-a51a-c93bbb6aa563"). InnerVolumeSpecName "kube-api-access-4g5hf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.999317 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6cb4aebe-be10-4574-8851-993dc42a3759-operator-scripts\") pod \"6cb4aebe-be10-4574-8851-993dc42a3759\" (UID: \"6cb4aebe-be10-4574-8851-993dc42a3759\") " Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.999372 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-config\") pod \"8552c885-2651-4424-b0d6-f75a5b7487dc\" (UID: \"8552c885-2651-4424-b0d6-f75a5b7487dc\") " Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.999422 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-dns-svc\") pod \"8552c885-2651-4424-b0d6-f75a5b7487dc\" (UID: \"8552c885-2651-4424-b0d6-f75a5b7487dc\") " Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.999503 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-ovsdbserver-nb\") pod \"8552c885-2651-4424-b0d6-f75a5b7487dc\" (UID: \"8552c885-2651-4424-b0d6-f75a5b7487dc\") " Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.999570 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2zfv\" (UniqueName: \"kubernetes.io/projected/6cb4aebe-be10-4574-8851-993dc42a3759-kube-api-access-h2zfv\") pod \"6cb4aebe-be10-4574-8851-993dc42a3759\" (UID: \"6cb4aebe-be10-4574-8851-993dc42a3759\") " Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.999635 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c2f74\" (UniqueName: \"kubernetes.io/projected/8552c885-2651-4424-b0d6-f75a5b7487dc-kube-api-access-c2f74\") pod \"8552c885-2651-4424-b0d6-f75a5b7487dc\" (UID: \"8552c885-2651-4424-b0d6-f75a5b7487dc\") " Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.999681 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-ovsdbserver-sb\") pod \"8552c885-2651-4424-b0d6-f75a5b7487dc\" (UID: \"8552c885-2651-4424-b0d6-f75a5b7487dc\") " Mar 17 09:30:43 crc kubenswrapper[4813]: I0317 09:30:43.999774 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6cb4aebe-be10-4574-8851-993dc42a3759-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6cb4aebe-be10-4574-8851-993dc42a3759" (UID: "6cb4aebe-be10-4574-8851-993dc42a3759"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.000062 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4g5hf\" (UniqueName: \"kubernetes.io/projected/a10880b9-b0fb-4510-a51a-c93bbb6aa563-kube-api-access-4g5hf\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.000084 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ae41df76-1944-4060-81a1-b28c7c839faf-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.000097 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bptw\" (UniqueName: \"kubernetes.io/projected/3666f515-3ed8-486a-a775-eea5c4463580-kube-api-access-4bptw\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.000110 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fkkh6\" (UniqueName: \"kubernetes.io/projected/ae41df76-1944-4060-81a1-b28c7c839faf-kube-api-access-fkkh6\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.000123 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7pndq\" (UniqueName: \"kubernetes.io/projected/5a047928-f15b-4c01-a4ef-b1c4b4d45cc9-kube-api-access-7pndq\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.000136 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6cb4aebe-be10-4574-8851-993dc42a3759-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.000148 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a10880b9-b0fb-4510-a51a-c93bbb6aa563-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.000161 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3666f515-3ed8-486a-a775-eea5c4463580-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.000174 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nm7v4\" (UniqueName: \"kubernetes.io/projected/34bc865e-b877-44b4-a57d-d7f390f60414-kube-api-access-nm7v4\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.000188 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/34bc865e-b877-44b4-a57d-d7f390f60414-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.000201 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5a047928-f15b-4c01-a4ef-b1c4b4d45cc9-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.003403 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6cb4aebe-be10-4574-8851-993dc42a3759-kube-api-access-h2zfv" (OuterVolumeSpecName: "kube-api-access-h2zfv") pod "6cb4aebe-be10-4574-8851-993dc42a3759" (UID: "6cb4aebe-be10-4574-8851-993dc42a3759"). InnerVolumeSpecName "kube-api-access-h2zfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.003868 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8552c885-2651-4424-b0d6-f75a5b7487dc-kube-api-access-c2f74" (OuterVolumeSpecName: "kube-api-access-c2f74") pod "8552c885-2651-4424-b0d6-f75a5b7487dc" (UID: "8552c885-2651-4424-b0d6-f75a5b7487dc"). InnerVolumeSpecName "kube-api-access-c2f74". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.037105 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8552c885-2651-4424-b0d6-f75a5b7487dc" (UID: "8552c885-2651-4424-b0d6-f75a5b7487dc"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.050625 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8552c885-2651-4424-b0d6-f75a5b7487dc" (UID: "8552c885-2651-4424-b0d6-f75a5b7487dc"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.074639 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8552c885-2651-4424-b0d6-f75a5b7487dc" (UID: "8552c885-2651-4424-b0d6-f75a5b7487dc"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.075147 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-config" (OuterVolumeSpecName: "config") pod "8552c885-2651-4424-b0d6-f75a5b7487dc" (UID: "8552c885-2651-4424-b0d6-f75a5b7487dc"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.102633 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c2f74\" (UniqueName: \"kubernetes.io/projected/8552c885-2651-4424-b0d6-f75a5b7487dc-kube-api-access-c2f74\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.102676 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.102689 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.102701 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.102713 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8552c885-2651-4424-b0d6-f75a5b7487dc-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.102728 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2zfv\" (UniqueName: \"kubernetes.io/projected/6cb4aebe-be10-4574-8851-993dc42a3759-kube-api-access-h2zfv\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.277986 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-8098-account-create-update-j4khm" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.278001 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-8098-account-create-update-j4khm" event={"ID":"a10880b9-b0fb-4510-a51a-c93bbb6aa563","Type":"ContainerDied","Data":"bdddcd7cb666fe4046e482824738cd03a67418539cb028a9186e53673ce67707"} Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.278059 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bdddcd7cb666fe4046e482824738cd03a67418539cb028a9186e53673ce67707" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.279570 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xlbs6" event={"ID":"f531bb05-ff10-4a9b-a8f4-58dc28a8007a","Type":"ContainerStarted","Data":"bd4a1f290d52c558f9005e068c1bc067598630436ad751660d444966ab779596"} Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.282007 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-e948-account-create-update-prpwx" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.282076 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-e948-account-create-update-prpwx" event={"ID":"3666f515-3ed8-486a-a775-eea5c4463580","Type":"ContainerDied","Data":"88f24b266ffaba1aa9cf9eb521c0817c0ab693b1bc3770f24f029f5a2287d51b"} Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.282135 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="88f24b266ffaba1aa9cf9eb521c0817c0ab693b1bc3770f24f029f5a2287d51b" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.296217 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-f43a-account-create-update-mxdk5" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.296204 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-f43a-account-create-update-mxdk5" event={"ID":"34bc865e-b877-44b4-a57d-d7f390f60414","Type":"ContainerDied","Data":"49df134cffcb0cf867f2189e839d8f0a5b5207e58e0319028fd99c6eb27dfcaa"} Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.296306 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49df134cffcb0cf867f2189e839d8f0a5b5207e58e0319028fd99c6eb27dfcaa" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.299036 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-pn9ft" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.299174 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-pn9ft" event={"ID":"5a047928-f15b-4c01-a4ef-b1c4b4d45cc9","Type":"ContainerDied","Data":"d353d5e812f81cdcf7741643f7148d3c13bbf5a360ed58cf303df722cdb60c27"} Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.299309 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d353d5e812f81cdcf7741643f7148d3c13bbf5a360ed58cf303df722cdb60c27" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.302028 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-79mzp" event={"ID":"6cb4aebe-be10-4574-8851-993dc42a3759","Type":"ContainerDied","Data":"a2393c81f55a914413980d991bfeb3490466d5a43f8dddb2b99cfa493dedb25f"} Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.302088 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2393c81f55a914413980d991bfeb3490466d5a43f8dddb2b99cfa493dedb25f" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.302090 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-79mzp" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.304781 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-b25vw" event={"ID":"8552c885-2651-4424-b0d6-f75a5b7487dc","Type":"ContainerDied","Data":"dd9b580dc06a35b506c8734f8284b9ced2e59287231eac3456b7472045d76499"} Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.304823 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-b25vw" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.304853 4813 scope.go:117] "RemoveContainer" containerID="0f2903c7e92e7817c2c04a7025fca74674207ad263f5ddacaf07b970622b3116" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.307150 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-xlbs6" podStartSLOduration=2.015291693 podStartE2EDuration="6.307130595s" podCreationTimestamp="2026-03-17 09:30:38 +0000 UTC" firstStartedPulling="2026-03-17 09:30:39.264248213 +0000 UTC m=+1261.365051712" lastFinishedPulling="2026-03-17 09:30:43.556087115 +0000 UTC m=+1265.656890614" observedRunningTime="2026-03-17 09:30:44.304978637 +0000 UTC m=+1266.405782176" watchObservedRunningTime="2026-03-17 09:30:44.307130595 +0000 UTC m=+1266.407934104" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.311918 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-b8kql" event={"ID":"ae41df76-1944-4060-81a1-b28c7c839faf","Type":"ContainerDied","Data":"589ad2a3e9b14dec62a01273af8a7676ac7c79955283508d520f0417fcc951f2"} Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.311976 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="589ad2a3e9b14dec62a01273af8a7676ac7c79955283508d520f0417fcc951f2" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.311987 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-b8kql" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.329842 4813 scope.go:117] "RemoveContainer" containerID="8cdd5e139009b415620c6c446d0d0577abceb05756f4f96280bcb870570be28e" Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.368023 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-b25vw"] Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.375047 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-b25vw"] Mar 17 09:30:44 crc kubenswrapper[4813]: I0317 09:30:44.739834 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8552c885-2651-4424-b0d6-f75a5b7487dc" path="/var/lib/kubelet/pods/8552c885-2651-4424-b0d6-f75a5b7487dc/volumes" Mar 17 09:30:47 crc kubenswrapper[4813]: I0317 09:30:47.361505 4813 generic.go:334] "Generic (PLEG): container finished" podID="f531bb05-ff10-4a9b-a8f4-58dc28a8007a" containerID="bd4a1f290d52c558f9005e068c1bc067598630436ad751660d444966ab779596" exitCode=0 Mar 17 09:30:47 crc kubenswrapper[4813]: I0317 09:30:47.361569 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xlbs6" event={"ID":"f531bb05-ff10-4a9b-a8f4-58dc28a8007a","Type":"ContainerDied","Data":"bd4a1f290d52c558f9005e068c1bc067598630436ad751660d444966ab779596"} Mar 17 09:30:48 crc kubenswrapper[4813]: I0317 09:30:48.751038 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xlbs6" Mar 17 09:30:48 crc kubenswrapper[4813]: I0317 09:30:48.790709 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-thlpr\" (UniqueName: \"kubernetes.io/projected/f531bb05-ff10-4a9b-a8f4-58dc28a8007a-kube-api-access-thlpr\") pod \"f531bb05-ff10-4a9b-a8f4-58dc28a8007a\" (UID: \"f531bb05-ff10-4a9b-a8f4-58dc28a8007a\") " Mar 17 09:30:48 crc kubenswrapper[4813]: I0317 09:30:48.790854 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f531bb05-ff10-4a9b-a8f4-58dc28a8007a-config-data\") pod \"f531bb05-ff10-4a9b-a8f4-58dc28a8007a\" (UID: \"f531bb05-ff10-4a9b-a8f4-58dc28a8007a\") " Mar 17 09:30:48 crc kubenswrapper[4813]: I0317 09:30:48.791672 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f531bb05-ff10-4a9b-a8f4-58dc28a8007a-combined-ca-bundle\") pod \"f531bb05-ff10-4a9b-a8f4-58dc28a8007a\" (UID: \"f531bb05-ff10-4a9b-a8f4-58dc28a8007a\") " Mar 17 09:30:48 crc kubenswrapper[4813]: I0317 09:30:48.796062 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f531bb05-ff10-4a9b-a8f4-58dc28a8007a-kube-api-access-thlpr" (OuterVolumeSpecName: "kube-api-access-thlpr") pod "f531bb05-ff10-4a9b-a8f4-58dc28a8007a" (UID: "f531bb05-ff10-4a9b-a8f4-58dc28a8007a"). InnerVolumeSpecName "kube-api-access-thlpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:48 crc kubenswrapper[4813]: I0317 09:30:48.813463 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f531bb05-ff10-4a9b-a8f4-58dc28a8007a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f531bb05-ff10-4a9b-a8f4-58dc28a8007a" (UID: "f531bb05-ff10-4a9b-a8f4-58dc28a8007a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:48 crc kubenswrapper[4813]: I0317 09:30:48.827980 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f531bb05-ff10-4a9b-a8f4-58dc28a8007a-config-data" (OuterVolumeSpecName: "config-data") pod "f531bb05-ff10-4a9b-a8f4-58dc28a8007a" (UID: "f531bb05-ff10-4a9b-a8f4-58dc28a8007a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:48 crc kubenswrapper[4813]: I0317 09:30:48.893471 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f531bb05-ff10-4a9b-a8f4-58dc28a8007a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:48 crc kubenswrapper[4813]: I0317 09:30:48.893509 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-thlpr\" (UniqueName: \"kubernetes.io/projected/f531bb05-ff10-4a9b-a8f4-58dc28a8007a-kube-api-access-thlpr\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:48 crc kubenswrapper[4813]: I0317 09:30:48.893524 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f531bb05-ff10-4a9b-a8f4-58dc28a8007a-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.382288 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-xlbs6" event={"ID":"f531bb05-ff10-4a9b-a8f4-58dc28a8007a","Type":"ContainerDied","Data":"8823cfcec655ecd7985551e2040e420b3d2ab43d411c25adeea3d32efa6b967a"} Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.382328 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8823cfcec655ecd7985551e2040e420b3d2ab43d411c25adeea3d32efa6b967a" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.382352 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-xlbs6" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.647913 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6"] Mar 17 09:30:49 crc kubenswrapper[4813]: E0317 09:30:49.656120 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f531bb05-ff10-4a9b-a8f4-58dc28a8007a" containerName="keystone-db-sync" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.656155 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f531bb05-ff10-4a9b-a8f4-58dc28a8007a" containerName="keystone-db-sync" Mar 17 09:30:49 crc kubenswrapper[4813]: E0317 09:30:49.656168 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a047928-f15b-4c01-a4ef-b1c4b4d45cc9" containerName="mariadb-database-create" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.656179 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a047928-f15b-4c01-a4ef-b1c4b4d45cc9" containerName="mariadb-database-create" Mar 17 09:30:49 crc kubenswrapper[4813]: E0317 09:30:49.656204 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="34bc865e-b877-44b4-a57d-d7f390f60414" containerName="mariadb-account-create-update" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.656213 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="34bc865e-b877-44b4-a57d-d7f390f60414" containerName="mariadb-account-create-update" Mar 17 09:30:49 crc kubenswrapper[4813]: E0317 09:30:49.656228 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3666f515-3ed8-486a-a775-eea5c4463580" containerName="mariadb-account-create-update" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.656235 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3666f515-3ed8-486a-a775-eea5c4463580" containerName="mariadb-account-create-update" Mar 17 09:30:49 crc kubenswrapper[4813]: E0317 09:30:49.656258 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6cb4aebe-be10-4574-8851-993dc42a3759" containerName="mariadb-database-create" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.656267 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6cb4aebe-be10-4574-8851-993dc42a3759" containerName="mariadb-database-create" Mar 17 09:30:49 crc kubenswrapper[4813]: E0317 09:30:49.656275 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae41df76-1944-4060-81a1-b28c7c839faf" containerName="mariadb-database-create" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.656284 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae41df76-1944-4060-81a1-b28c7c839faf" containerName="mariadb-database-create" Mar 17 09:30:49 crc kubenswrapper[4813]: E0317 09:30:49.656299 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a10880b9-b0fb-4510-a51a-c93bbb6aa563" containerName="mariadb-account-create-update" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.656307 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a10880b9-b0fb-4510-a51a-c93bbb6aa563" containerName="mariadb-account-create-update" Mar 17 09:30:49 crc kubenswrapper[4813]: E0317 09:30:49.656326 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8552c885-2651-4424-b0d6-f75a5b7487dc" containerName="dnsmasq-dns" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.656334 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8552c885-2651-4424-b0d6-f75a5b7487dc" containerName="dnsmasq-dns" Mar 17 09:30:49 crc kubenswrapper[4813]: E0317 09:30:49.656346 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8552c885-2651-4424-b0d6-f75a5b7487dc" containerName="init" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.656353 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8552c885-2651-4424-b0d6-f75a5b7487dc" containerName="init" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.656570 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a047928-f15b-4c01-a4ef-b1c4b4d45cc9" containerName="mariadb-database-create" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.656590 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a10880b9-b0fb-4510-a51a-c93bbb6aa563" containerName="mariadb-account-create-update" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.656628 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8552c885-2651-4424-b0d6-f75a5b7487dc" containerName="dnsmasq-dns" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.656645 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="3666f515-3ed8-486a-a775-eea5c4463580" containerName="mariadb-account-create-update" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.656654 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae41df76-1944-4060-81a1-b28c7c839faf" containerName="mariadb-database-create" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.656665 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="34bc865e-b877-44b4-a57d-d7f390f60414" containerName="mariadb-account-create-update" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.656676 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="6cb4aebe-be10-4574-8851-993dc42a3759" containerName="mariadb-database-create" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.656686 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f531bb05-ff10-4a9b-a8f4-58dc28a8007a" containerName="keystone-db-sync" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.657706 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.671473 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6"] Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.706342 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-config\") pod \"dnsmasq-dns-5c5cc7c5ff-cfnp6\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.706463 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-cfnp6\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.706530 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-cfnp6\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.706561 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-cfnp6\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.706639 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trs7m\" (UniqueName: \"kubernetes.io/projected/cbe63abb-2664-4824-9f02-001b8ec06edb-kube-api-access-trs7m\") pod \"dnsmasq-dns-5c5cc7c5ff-cfnp6\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.706662 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-cfnp6\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.735899 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-p85pk"] Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.736978 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.743775 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.744027 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.744159 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-tnvpg" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.744207 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.747134 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-p85pk"] Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.759824 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.814769 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-cfnp6\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.814921 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-cfnp6\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.814953 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-config-data\") pod \"keystone-bootstrap-p85pk\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.814982 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-cfnp6\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.815045 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-scripts\") pod \"keystone-bootstrap-p85pk\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.815149 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-combined-ca-bundle\") pod \"keystone-bootstrap-p85pk\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.815232 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trs7m\" (UniqueName: \"kubernetes.io/projected/cbe63abb-2664-4824-9f02-001b8ec06edb-kube-api-access-trs7m\") pod \"dnsmasq-dns-5c5cc7c5ff-cfnp6\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.815258 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-cfnp6\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.815338 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-config\") pod \"dnsmasq-dns-5c5cc7c5ff-cfnp6\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.815368 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-credential-keys\") pod \"keystone-bootstrap-p85pk\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.815407 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z9xkb\" (UniqueName: \"kubernetes.io/projected/2728d43b-fbec-4573-9f2f-5d29b57e82d6-kube-api-access-z9xkb\") pod \"keystone-bootstrap-p85pk\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.815491 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-fernet-keys\") pod \"keystone-bootstrap-p85pk\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.815564 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-dns-svc\") pod \"dnsmasq-dns-5c5cc7c5ff-cfnp6\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.816293 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-dns-swift-storage-0\") pod \"dnsmasq-dns-5c5cc7c5ff-cfnp6\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.816715 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-ovsdbserver-nb\") pod \"dnsmasq-dns-5c5cc7c5ff-cfnp6\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.817561 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-ovsdbserver-sb\") pod \"dnsmasq-dns-5c5cc7c5ff-cfnp6\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.817964 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-config\") pod \"dnsmasq-dns-5c5cc7c5ff-cfnp6\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.861390 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trs7m\" (UniqueName: \"kubernetes.io/projected/cbe63abb-2664-4824-9f02-001b8ec06edb-kube-api-access-trs7m\") pod \"dnsmasq-dns-5c5cc7c5ff-cfnp6\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.887989 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7fdb788585-bnbmf"] Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.889390 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fdb788585-bnbmf" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.896399 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-fnhk9" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.896570 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.896703 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.896813 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.921830 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-config-data\") pod \"keystone-bootstrap-p85pk\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.921879 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-scripts\") pod \"keystone-bootstrap-p85pk\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.921903 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/93e11613-c787-402d-aabb-384aedd72deb-scripts\") pod \"horizon-7fdb788585-bnbmf\" (UID: \"93e11613-c787-402d-aabb-384aedd72deb\") " pod="openstack/horizon-7fdb788585-bnbmf" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.921922 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-combined-ca-bundle\") pod \"keystone-bootstrap-p85pk\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.921938 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93e11613-c787-402d-aabb-384aedd72deb-logs\") pod \"horizon-7fdb788585-bnbmf\" (UID: \"93e11613-c787-402d-aabb-384aedd72deb\") " pod="openstack/horizon-7fdb788585-bnbmf" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.921959 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbx5x\" (UniqueName: \"kubernetes.io/projected/93e11613-c787-402d-aabb-384aedd72deb-kube-api-access-qbx5x\") pod \"horizon-7fdb788585-bnbmf\" (UID: \"93e11613-c787-402d-aabb-384aedd72deb\") " pod="openstack/horizon-7fdb788585-bnbmf" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.921995 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-credential-keys\") pod \"keystone-bootstrap-p85pk\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.922012 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93e11613-c787-402d-aabb-384aedd72deb-config-data\") pod \"horizon-7fdb788585-bnbmf\" (UID: \"93e11613-c787-402d-aabb-384aedd72deb\") " pod="openstack/horizon-7fdb788585-bnbmf" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.922030 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z9xkb\" (UniqueName: \"kubernetes.io/projected/2728d43b-fbec-4573-9f2f-5d29b57e82d6-kube-api-access-z9xkb\") pod \"keystone-bootstrap-p85pk\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.922052 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-fernet-keys\") pod \"keystone-bootstrap-p85pk\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.922075 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/93e11613-c787-402d-aabb-384aedd72deb-horizon-secret-key\") pod \"horizon-7fdb788585-bnbmf\" (UID: \"93e11613-c787-402d-aabb-384aedd72deb\") " pod="openstack/horizon-7fdb788585-bnbmf" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.937098 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-credential-keys\") pod \"keystone-bootstrap-p85pk\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.939704 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-combined-ca-bundle\") pod \"keystone-bootstrap-p85pk\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.944732 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-fernet-keys\") pod \"keystone-bootstrap-p85pk\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.945243 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-config-data\") pod \"keystone-bootstrap-p85pk\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.945340 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-jtbdz"] Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.946347 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jtbdz" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.948236 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-scripts\") pod \"keystone-bootstrap-p85pk\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.950107 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7fdb788585-bnbmf"] Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.952849 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.953022 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-js8px" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.953189 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.966785 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z9xkb\" (UniqueName: \"kubernetes.io/projected/2728d43b-fbec-4573-9f2f-5d29b57e82d6-kube-api-access-z9xkb\") pod \"keystone-bootstrap-p85pk\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.971954 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-jtbdz"] Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.986953 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-27g9l"] Mar 17 09:30:49 crc kubenswrapper[4813]: I0317 09:30:49.987868 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-27g9l" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:49.998070 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:49.998717 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-6rfxv" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:49.999953 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.010309 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.027232 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-db-sync-config-data\") pod \"cinder-db-sync-27g9l\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " pod="openstack/cinder-db-sync-27g9l" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.027295 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/765fce5e-7101-42a5-8ad1-172c967fd2c3-combined-ca-bundle\") pod \"neutron-db-sync-jtbdz\" (UID: \"765fce5e-7101-42a5-8ad1-172c967fd2c3\") " pod="openstack/neutron-db-sync-jtbdz" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.027331 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zfpnh\" (UniqueName: \"kubernetes.io/projected/41809604-4a36-4453-b381-66b4c5cf1c43-kube-api-access-zfpnh\") pod \"cinder-db-sync-27g9l\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " pod="openstack/cinder-db-sync-27g9l" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.027365 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/765fce5e-7101-42a5-8ad1-172c967fd2c3-config\") pod \"neutron-db-sync-jtbdz\" (UID: \"765fce5e-7101-42a5-8ad1-172c967fd2c3\") " pod="openstack/neutron-db-sync-jtbdz" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.027393 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/93e11613-c787-402d-aabb-384aedd72deb-scripts\") pod \"horizon-7fdb788585-bnbmf\" (UID: \"93e11613-c787-402d-aabb-384aedd72deb\") " pod="openstack/horizon-7fdb788585-bnbmf" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.027411 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41809604-4a36-4453-b381-66b4c5cf1c43-etc-machine-id\") pod \"cinder-db-sync-27g9l\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " pod="openstack/cinder-db-sync-27g9l" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.027428 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93e11613-c787-402d-aabb-384aedd72deb-logs\") pod \"horizon-7fdb788585-bnbmf\" (UID: \"93e11613-c787-402d-aabb-384aedd72deb\") " pod="openstack/horizon-7fdb788585-bnbmf" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.027448 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbx5x\" (UniqueName: \"kubernetes.io/projected/93e11613-c787-402d-aabb-384aedd72deb-kube-api-access-qbx5x\") pod \"horizon-7fdb788585-bnbmf\" (UID: \"93e11613-c787-402d-aabb-384aedd72deb\") " pod="openstack/horizon-7fdb788585-bnbmf" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.027491 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-scripts\") pod \"cinder-db-sync-27g9l\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " pod="openstack/cinder-db-sync-27g9l" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.027515 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93e11613-c787-402d-aabb-384aedd72deb-config-data\") pod \"horizon-7fdb788585-bnbmf\" (UID: \"93e11613-c787-402d-aabb-384aedd72deb\") " pod="openstack/horizon-7fdb788585-bnbmf" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.027546 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/93e11613-c787-402d-aabb-384aedd72deb-horizon-secret-key\") pod \"horizon-7fdb788585-bnbmf\" (UID: \"93e11613-c787-402d-aabb-384aedd72deb\") " pod="openstack/horizon-7fdb788585-bnbmf" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.027563 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5q4bp\" (UniqueName: \"kubernetes.io/projected/765fce5e-7101-42a5-8ad1-172c967fd2c3-kube-api-access-5q4bp\") pod \"neutron-db-sync-jtbdz\" (UID: \"765fce5e-7101-42a5-8ad1-172c967fd2c3\") " pod="openstack/neutron-db-sync-jtbdz" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.027582 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-config-data\") pod \"cinder-db-sync-27g9l\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " pod="openstack/cinder-db-sync-27g9l" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.029688 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-27g9l"] Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.031765 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-combined-ca-bundle\") pod \"cinder-db-sync-27g9l\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " pod="openstack/cinder-db-sync-27g9l" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.032135 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93e11613-c787-402d-aabb-384aedd72deb-logs\") pod \"horizon-7fdb788585-bnbmf\" (UID: \"93e11613-c787-402d-aabb-384aedd72deb\") " pod="openstack/horizon-7fdb788585-bnbmf" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.044300 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/93e11613-c787-402d-aabb-384aedd72deb-scripts\") pod \"horizon-7fdb788585-bnbmf\" (UID: \"93e11613-c787-402d-aabb-384aedd72deb\") " pod="openstack/horizon-7fdb788585-bnbmf" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.049697 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93e11613-c787-402d-aabb-384aedd72deb-config-data\") pod \"horizon-7fdb788585-bnbmf\" (UID: \"93e11613-c787-402d-aabb-384aedd72deb\") " pod="openstack/horizon-7fdb788585-bnbmf" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.059741 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/93e11613-c787-402d-aabb-384aedd72deb-horizon-secret-key\") pod \"horizon-7fdb788585-bnbmf\" (UID: \"93e11613-c787-402d-aabb-384aedd72deb\") " pod="openstack/horizon-7fdb788585-bnbmf" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.068179 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.080128 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.080197 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbx5x\" (UniqueName: \"kubernetes.io/projected/93e11613-c787-402d-aabb-384aedd72deb-kube-api-access-qbx5x\") pod \"horizon-7fdb788585-bnbmf\" (UID: \"93e11613-c787-402d-aabb-384aedd72deb\") " pod="openstack/horizon-7fdb788585-bnbmf" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.086907 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.092068 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.092265 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.115121 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.167349 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c7k8\" (UniqueName: \"kubernetes.io/projected/d927c167-348f-47f8-9860-b21dac3ca309-kube-api-access-4c7k8\") pod \"ceilometer-0\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.167433 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/765fce5e-7101-42a5-8ad1-172c967fd2c3-combined-ca-bundle\") pod \"neutron-db-sync-jtbdz\" (UID: \"765fce5e-7101-42a5-8ad1-172c967fd2c3\") " pod="openstack/neutron-db-sync-jtbdz" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.167473 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.167559 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zfpnh\" (UniqueName: \"kubernetes.io/projected/41809604-4a36-4453-b381-66b4c5cf1c43-kube-api-access-zfpnh\") pod \"cinder-db-sync-27g9l\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " pod="openstack/cinder-db-sync-27g9l" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.167628 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.167702 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/765fce5e-7101-42a5-8ad1-172c967fd2c3-config\") pod \"neutron-db-sync-jtbdz\" (UID: \"765fce5e-7101-42a5-8ad1-172c967fd2c3\") " pod="openstack/neutron-db-sync-jtbdz" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.167751 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41809604-4a36-4453-b381-66b4c5cf1c43-etc-machine-id\") pod \"cinder-db-sync-27g9l\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " pod="openstack/cinder-db-sync-27g9l" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.167846 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-scripts\") pod \"cinder-db-sync-27g9l\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " pod="openstack/cinder-db-sync-27g9l" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.167904 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d927c167-348f-47f8-9860-b21dac3ca309-log-httpd\") pod \"ceilometer-0\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.167977 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-config-data\") pod \"ceilometer-0\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.168014 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5q4bp\" (UniqueName: \"kubernetes.io/projected/765fce5e-7101-42a5-8ad1-172c967fd2c3-kube-api-access-5q4bp\") pod \"neutron-db-sync-jtbdz\" (UID: \"765fce5e-7101-42a5-8ad1-172c967fd2c3\") " pod="openstack/neutron-db-sync-jtbdz" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.168054 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-scripts\") pod \"ceilometer-0\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.168086 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-config-data\") pod \"cinder-db-sync-27g9l\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " pod="openstack/cinder-db-sync-27g9l" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.168115 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-combined-ca-bundle\") pod \"cinder-db-sync-27g9l\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " pod="openstack/cinder-db-sync-27g9l" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.168149 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d927c167-348f-47f8-9860-b21dac3ca309-run-httpd\") pod \"ceilometer-0\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.168210 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-db-sync-config-data\") pod \"cinder-db-sync-27g9l\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " pod="openstack/cinder-db-sync-27g9l" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.183836 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41809604-4a36-4453-b381-66b4c5cf1c43-etc-machine-id\") pod \"cinder-db-sync-27g9l\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " pod="openstack/cinder-db-sync-27g9l" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.184383 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-8654c79df-tc6h4"] Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.201884 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/765fce5e-7101-42a5-8ad1-172c967fd2c3-config\") pod \"neutron-db-sync-jtbdz\" (UID: \"765fce5e-7101-42a5-8ad1-172c967fd2c3\") " pod="openstack/neutron-db-sync-jtbdz" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.207799 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8654c79df-tc6h4" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.224112 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fdb788585-bnbmf" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.231891 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-mxghx"] Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.238859 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-mxghx" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.239989 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/765fce5e-7101-42a5-8ad1-172c967fd2c3-combined-ca-bundle\") pod \"neutron-db-sync-jtbdz\" (UID: \"765fce5e-7101-42a5-8ad1-172c967fd2c3\") " pod="openstack/neutron-db-sync-jtbdz" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.243945 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-kkzw4" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.244427 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.250047 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5q4bp\" (UniqueName: \"kubernetes.io/projected/765fce5e-7101-42a5-8ad1-172c967fd2c3-kube-api-access-5q4bp\") pod \"neutron-db-sync-jtbdz\" (UID: \"765fce5e-7101-42a5-8ad1-172c967fd2c3\") " pod="openstack/neutron-db-sync-jtbdz" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.252080 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-db-sync-config-data\") pod \"cinder-db-sync-27g9l\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " pod="openstack/cinder-db-sync-27g9l" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.253508 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6"] Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.259346 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-combined-ca-bundle\") pod \"cinder-db-sync-27g9l\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " pod="openstack/cinder-db-sync-27g9l" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.272127 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-mxghx"] Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.285986 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-config-data\") pod \"cinder-db-sync-27g9l\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " pod="openstack/cinder-db-sync-27g9l" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.286999 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-scripts\") pod \"cinder-db-sync-27g9l\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " pod="openstack/cinder-db-sync-27g9l" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.289053 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d927c167-348f-47f8-9860-b21dac3ca309-run-httpd\") pod \"ceilometer-0\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.289171 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c7k8\" (UniqueName: \"kubernetes.io/projected/d927c167-348f-47f8-9860-b21dac3ca309-kube-api-access-4c7k8\") pod \"ceilometer-0\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.289239 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.289316 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-horizon-secret-key\") pod \"horizon-8654c79df-tc6h4\" (UID: \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\") " pod="openstack/horizon-8654c79df-tc6h4" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.289338 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.289440 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-logs\") pod \"horizon-8654c79df-tc6h4\" (UID: \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\") " pod="openstack/horizon-8654c79df-tc6h4" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.289507 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h786\" (UniqueName: \"kubernetes.io/projected/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-kube-api-access-9h786\") pod \"horizon-8654c79df-tc6h4\" (UID: \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\") " pod="openstack/horizon-8654c79df-tc6h4" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.289540 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-scripts\") pod \"horizon-8654c79df-tc6h4\" (UID: \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\") " pod="openstack/horizon-8654c79df-tc6h4" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.289564 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d927c167-348f-47f8-9860-b21dac3ca309-log-httpd\") pod \"ceilometer-0\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.289610 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-config-data\") pod \"horizon-8654c79df-tc6h4\" (UID: \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\") " pod="openstack/horizon-8654c79df-tc6h4" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.289640 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-config-data\") pod \"ceilometer-0\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.289669 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-scripts\") pod \"ceilometer-0\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.290581 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d927c167-348f-47f8-9860-b21dac3ca309-log-httpd\") pod \"ceilometer-0\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.290767 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d927c167-348f-47f8-9860-b21dac3ca309-run-httpd\") pod \"ceilometer-0\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.296855 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-config-data\") pod \"ceilometer-0\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.302045 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zfpnh\" (UniqueName: \"kubernetes.io/projected/41809604-4a36-4453-b381-66b4c5cf1c43-kube-api-access-zfpnh\") pod \"cinder-db-sync-27g9l\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " pod="openstack/cinder-db-sync-27g9l" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.309102 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.318338 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-scripts\") pod \"ceilometer-0\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.318696 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.327434 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c7k8\" (UniqueName: \"kubernetes.io/projected/d927c167-348f-47f8-9860-b21dac3ca309-kube-api-access-4c7k8\") pod \"ceilometer-0\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.344356 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8654c79df-tc6h4"] Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.375810 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-hkmc6"] Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.377290 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hkmc6" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.384103 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-5l9nj" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.395923 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.396107 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.410296 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c6cd309-922e-4be6-b2d9-12d2aca54b78-combined-ca-bundle\") pod \"barbican-db-sync-mxghx\" (UID: \"5c6cd309-922e-4be6-b2d9-12d2aca54b78\") " pod="openstack/barbican-db-sync-mxghx" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.410427 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-horizon-secret-key\") pod \"horizon-8654c79df-tc6h4\" (UID: \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\") " pod="openstack/horizon-8654c79df-tc6h4" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.410492 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-logs\") pod \"horizon-8654c79df-tc6h4\" (UID: \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\") " pod="openstack/horizon-8654c79df-tc6h4" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.410535 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h786\" (UniqueName: \"kubernetes.io/projected/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-kube-api-access-9h786\") pod \"horizon-8654c79df-tc6h4\" (UID: \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\") " pod="openstack/horizon-8654c79df-tc6h4" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.410568 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lgwjb\" (UniqueName: \"kubernetes.io/projected/5c6cd309-922e-4be6-b2d9-12d2aca54b78-kube-api-access-lgwjb\") pod \"barbican-db-sync-mxghx\" (UID: \"5c6cd309-922e-4be6-b2d9-12d2aca54b78\") " pod="openstack/barbican-db-sync-mxghx" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.410607 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-scripts\") pod \"horizon-8654c79df-tc6h4\" (UID: \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\") " pod="openstack/horizon-8654c79df-tc6h4" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.410637 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-config-data\") pod \"horizon-8654c79df-tc6h4\" (UID: \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\") " pod="openstack/horizon-8654c79df-tc6h4" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.410655 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5c6cd309-922e-4be6-b2d9-12d2aca54b78-db-sync-config-data\") pod \"barbican-db-sync-mxghx\" (UID: \"5c6cd309-922e-4be6-b2d9-12d2aca54b78\") " pod="openstack/barbican-db-sync-mxghx" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.411127 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-logs\") pod \"horizon-8654c79df-tc6h4\" (UID: \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\") " pod="openstack/horizon-8654c79df-tc6h4" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.412334 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-scripts\") pod \"horizon-8654c79df-tc6h4\" (UID: \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\") " pod="openstack/horizon-8654c79df-tc6h4" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.423636 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jtbdz" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.424609 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-config-data\") pod \"horizon-8654c79df-tc6h4\" (UID: \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\") " pod="openstack/horizon-8654c79df-tc6h4" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.425047 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-horizon-secret-key\") pod \"horizon-8654c79df-tc6h4\" (UID: \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\") " pod="openstack/horizon-8654c79df-tc6h4" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.445954 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-6wc77"] Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.448761 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h786\" (UniqueName: \"kubernetes.io/projected/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-kube-api-access-9h786\") pod \"horizon-8654c79df-tc6h4\" (UID: \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\") " pod="openstack/horizon-8654c79df-tc6h4" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.448932 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.495899 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-6wc77"] Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.509293 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.516633 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5c6cd309-922e-4be6-b2d9-12d2aca54b78-db-sync-config-data\") pod \"barbican-db-sync-mxghx\" (UID: \"5c6cd309-922e-4be6-b2d9-12d2aca54b78\") " pod="openstack/barbican-db-sync-mxghx" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.516675 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g5vfq\" (UniqueName: \"kubernetes.io/projected/1596452f-363d-4d92-82de-f96f1f83993a-kube-api-access-g5vfq\") pod \"dnsmasq-dns-8b5c85b87-6wc77\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.516700 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-6wc77\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.516718 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-6wc77\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.516742 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-6wc77\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.516765 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c6cd309-922e-4be6-b2d9-12d2aca54b78-combined-ca-bundle\") pod \"barbican-db-sync-mxghx\" (UID: \"5c6cd309-922e-4be6-b2d9-12d2aca54b78\") " pod="openstack/barbican-db-sync-mxghx" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.516783 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c613a9b-36b2-437d-9c11-252d7d8ce910-logs\") pod \"placement-db-sync-hkmc6\" (UID: \"9c613a9b-36b2-437d-9c11-252d7d8ce910\") " pod="openstack/placement-db-sync-hkmc6" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.522855 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-config\") pod \"dnsmasq-dns-8b5c85b87-6wc77\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.522968 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-6wc77\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.523081 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c613a9b-36b2-437d-9c11-252d7d8ce910-combined-ca-bundle\") pod \"placement-db-sync-hkmc6\" (UID: \"9c613a9b-36b2-437d-9c11-252d7d8ce910\") " pod="openstack/placement-db-sync-hkmc6" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.523107 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c613a9b-36b2-437d-9c11-252d7d8ce910-config-data\") pod \"placement-db-sync-hkmc6\" (UID: \"9c613a9b-36b2-437d-9c11-252d7d8ce910\") " pod="openstack/placement-db-sync-hkmc6" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.523132 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c613a9b-36b2-437d-9c11-252d7d8ce910-scripts\") pod \"placement-db-sync-hkmc6\" (UID: \"9c613a9b-36b2-437d-9c11-252d7d8ce910\") " pod="openstack/placement-db-sync-hkmc6" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.523157 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tnqs\" (UniqueName: \"kubernetes.io/projected/9c613a9b-36b2-437d-9c11-252d7d8ce910-kube-api-access-2tnqs\") pod \"placement-db-sync-hkmc6\" (UID: \"9c613a9b-36b2-437d-9c11-252d7d8ce910\") " pod="openstack/placement-db-sync-hkmc6" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.523181 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lgwjb\" (UniqueName: \"kubernetes.io/projected/5c6cd309-922e-4be6-b2d9-12d2aca54b78-kube-api-access-lgwjb\") pod \"barbican-db-sync-mxghx\" (UID: \"5c6cd309-922e-4be6-b2d9-12d2aca54b78\") " pod="openstack/barbican-db-sync-mxghx" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.526410 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-hkmc6"] Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.532389 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c6cd309-922e-4be6-b2d9-12d2aca54b78-combined-ca-bundle\") pod \"barbican-db-sync-mxghx\" (UID: \"5c6cd309-922e-4be6-b2d9-12d2aca54b78\") " pod="openstack/barbican-db-sync-mxghx" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.542730 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8654c79df-tc6h4" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.543556 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5c6cd309-922e-4be6-b2d9-12d2aca54b78-db-sync-config-data\") pod \"barbican-db-sync-mxghx\" (UID: \"5c6cd309-922e-4be6-b2d9-12d2aca54b78\") " pod="openstack/barbican-db-sync-mxghx" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.550988 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-27g9l" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.551501 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lgwjb\" (UniqueName: \"kubernetes.io/projected/5c6cd309-922e-4be6-b2d9-12d2aca54b78-kube-api-access-lgwjb\") pod \"barbican-db-sync-mxghx\" (UID: \"5c6cd309-922e-4be6-b2d9-12d2aca54b78\") " pod="openstack/barbican-db-sync-mxghx" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.594697 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.596116 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.597235 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-mxghx" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.601751 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.601987 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-bw5js" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.602128 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.602239 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.645386 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-6wc77\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.645774 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c613a9b-36b2-437d-9c11-252d7d8ce910-logs\") pod \"placement-db-sync-hkmc6\" (UID: \"9c613a9b-36b2-437d-9c11-252d7d8ce910\") " pod="openstack/placement-db-sync-hkmc6" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.645830 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-config\") pod \"dnsmasq-dns-8b5c85b87-6wc77\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.645876 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-6wc77\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.645932 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c613a9b-36b2-437d-9c11-252d7d8ce910-combined-ca-bundle\") pod \"placement-db-sync-hkmc6\" (UID: \"9c613a9b-36b2-437d-9c11-252d7d8ce910\") " pod="openstack/placement-db-sync-hkmc6" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.645954 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c613a9b-36b2-437d-9c11-252d7d8ce910-config-data\") pod \"placement-db-sync-hkmc6\" (UID: \"9c613a9b-36b2-437d-9c11-252d7d8ce910\") " pod="openstack/placement-db-sync-hkmc6" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.645978 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c613a9b-36b2-437d-9c11-252d7d8ce910-scripts\") pod \"placement-db-sync-hkmc6\" (UID: \"9c613a9b-36b2-437d-9c11-252d7d8ce910\") " pod="openstack/placement-db-sync-hkmc6" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.646000 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tnqs\" (UniqueName: \"kubernetes.io/projected/9c613a9b-36b2-437d-9c11-252d7d8ce910-kube-api-access-2tnqs\") pod \"placement-db-sync-hkmc6\" (UID: \"9c613a9b-36b2-437d-9c11-252d7d8ce910\") " pod="openstack/placement-db-sync-hkmc6" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.646039 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g5vfq\" (UniqueName: \"kubernetes.io/projected/1596452f-363d-4d92-82de-f96f1f83993a-kube-api-access-g5vfq\") pod \"dnsmasq-dns-8b5c85b87-6wc77\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.646071 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-6wc77\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.646093 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-6wc77\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.647000 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-ovsdbserver-nb\") pod \"dnsmasq-dns-8b5c85b87-6wc77\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.650543 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-dns-swift-storage-0\") pod \"dnsmasq-dns-8b5c85b87-6wc77\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.650944 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c613a9b-36b2-437d-9c11-252d7d8ce910-logs\") pod \"placement-db-sync-hkmc6\" (UID: \"9c613a9b-36b2-437d-9c11-252d7d8ce910\") " pod="openstack/placement-db-sync-hkmc6" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.651765 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.651844 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-dns-svc\") pod \"dnsmasq-dns-8b5c85b87-6wc77\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.652459 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-config\") pod \"dnsmasq-dns-8b5c85b87-6wc77\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.652474 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-ovsdbserver-sb\") pod \"dnsmasq-dns-8b5c85b87-6wc77\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.659498 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c613a9b-36b2-437d-9c11-252d7d8ce910-combined-ca-bundle\") pod \"placement-db-sync-hkmc6\" (UID: \"9c613a9b-36b2-437d-9c11-252d7d8ce910\") " pod="openstack/placement-db-sync-hkmc6" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.659571 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c613a9b-36b2-437d-9c11-252d7d8ce910-scripts\") pod \"placement-db-sync-hkmc6\" (UID: \"9c613a9b-36b2-437d-9c11-252d7d8ce910\") " pod="openstack/placement-db-sync-hkmc6" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.665512 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c613a9b-36b2-437d-9c11-252d7d8ce910-config-data\") pod \"placement-db-sync-hkmc6\" (UID: \"9c613a9b-36b2-437d-9c11-252d7d8ce910\") " pod="openstack/placement-db-sync-hkmc6" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.679249 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g5vfq\" (UniqueName: \"kubernetes.io/projected/1596452f-363d-4d92-82de-f96f1f83993a-kube-api-access-g5vfq\") pod \"dnsmasq-dns-8b5c85b87-6wc77\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.680374 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tnqs\" (UniqueName: \"kubernetes.io/projected/9c613a9b-36b2-437d-9c11-252d7d8ce910-kube-api-access-2tnqs\") pod \"placement-db-sync-hkmc6\" (UID: \"9c613a9b-36b2-437d-9c11-252d7d8ce910\") " pod="openstack/placement-db-sync-hkmc6" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.716520 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.718354 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.723474 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.723690 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.749469 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.749513 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b979q\" (UniqueName: \"kubernetes.io/projected/b5357557-3288-455b-8997-231364e39026-kube-api-access-b979q\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.749551 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5357557-3288-455b-8997-231364e39026-logs\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.749581 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b5357557-3288-455b-8997-231364e39026-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.749660 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.749695 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.749725 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-scripts\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.749752 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-config-data\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.755510 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hkmc6" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.808589 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.850993 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.851049 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6"] Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.851065 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-p85pk"] Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.852006 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5357557-3288-455b-8997-231364e39026-logs\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.852060 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b5357557-3288-455b-8997-231364e39026-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.852089 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7gc2g\" (UniqueName: \"kubernetes.io/projected/74a2cdb9-0725-4445-ba2a-4889c8c6a680-kube-api-access-7gc2g\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.852122 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.852154 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.852189 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-config-data\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.852223 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.852271 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.852317 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74a2cdb9-0725-4445-ba2a-4889c8c6a680-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.852341 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-scripts\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.852368 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.852393 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-scripts\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.852429 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-config-data\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.852459 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74a2cdb9-0725-4445-ba2a-4889c8c6a680-logs\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.852484 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.852513 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b979q\" (UniqueName: \"kubernetes.io/projected/b5357557-3288-455b-8997-231364e39026-kube-api-access-b979q\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.856717 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5357557-3288-455b-8997-231364e39026-logs\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.860733 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b5357557-3288-455b-8997-231364e39026-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.867925 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.868831 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.870281 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-config-data\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.873986 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-scripts\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: W0317 09:30:50.909651 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcbe63abb_2664_4824_9f02_001b8ec06edb.slice/crio-9a257ddace2c4e3501802f0370f9ae08b1e317ec5a3af9863d6206629b0b60c9 WatchSource:0}: Error finding container 9a257ddace2c4e3501802f0370f9ae08b1e317ec5a3af9863d6206629b0b60c9: Status 404 returned error can't find the container with id 9a257ddace2c4e3501802f0370f9ae08b1e317ec5a3af9863d6206629b0b60c9 Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.910501 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b979q\" (UniqueName: \"kubernetes.io/projected/b5357557-3288-455b-8997-231364e39026-kube-api-access-b979q\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.948265 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.953878 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74a2cdb9-0725-4445-ba2a-4889c8c6a680-logs\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.954068 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7gc2g\" (UniqueName: \"kubernetes.io/projected/74a2cdb9-0725-4445-ba2a-4889c8c6a680-kube-api-access-7gc2g\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.954187 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.954271 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.954371 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-config-data\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.954437 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74a2cdb9-0725-4445-ba2a-4889c8c6a680-logs\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.954561 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74a2cdb9-0725-4445-ba2a-4889c8c6a680-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.954655 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-scripts\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.954724 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.954734 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.959724 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74a2cdb9-0725-4445-ba2a-4889c8c6a680-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.975756 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.989504 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.992718 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-scripts\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:50 crc kubenswrapper[4813]: I0317 09:30:50.999751 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-config-data\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:51 crc kubenswrapper[4813]: I0317 09:30:51.003906 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7gc2g\" (UniqueName: \"kubernetes.io/projected/74a2cdb9-0725-4445-ba2a-4889c8c6a680-kube-api-access-7gc2g\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:51 crc kubenswrapper[4813]: I0317 09:30:51.006909 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:51 crc kubenswrapper[4813]: I0317 09:30:51.007530 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:51 crc kubenswrapper[4813]: I0317 09:30:51.245116 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 09:30:51 crc kubenswrapper[4813]: I0317 09:30:51.256669 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 09:30:51 crc kubenswrapper[4813]: I0317 09:30:51.477646 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-p85pk" event={"ID":"2728d43b-fbec-4573-9f2f-5d29b57e82d6","Type":"ContainerStarted","Data":"0915e99e93df5119ef52f738f246dcf25eb4dd780225e4b1f659ae48abeb232b"} Mar 17 09:30:51 crc kubenswrapper[4813]: I0317 09:30:51.478086 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-p85pk" event={"ID":"2728d43b-fbec-4573-9f2f-5d29b57e82d6","Type":"ContainerStarted","Data":"b49e32ad5b9b4a883def4c563e572945b93adac12f9225d90ec113d7790d0fe1"} Mar 17 09:30:51 crc kubenswrapper[4813]: I0317 09:30:51.494817 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" event={"ID":"cbe63abb-2664-4824-9f02-001b8ec06edb","Type":"ContainerStarted","Data":"9a257ddace2c4e3501802f0370f9ae08b1e317ec5a3af9863d6206629b0b60c9"} Mar 17 09:30:51 crc kubenswrapper[4813]: I0317 09:30:51.514832 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-p85pk" podStartSLOduration=2.514811472 podStartE2EDuration="2.514811472s" podCreationTimestamp="2026-03-17 09:30:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:30:51.501635637 +0000 UTC m=+1273.602439136" watchObservedRunningTime="2026-03-17 09:30:51.514811472 +0000 UTC m=+1273.615614971" Mar 17 09:30:51 crc kubenswrapper[4813]: W0317 09:30:51.637040 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod93e11613_c787_402d_aabb_384aedd72deb.slice/crio-b5597f7e16c010f932748244d694a6afa0a68d54c263e20549f6f52e31ad6b42 WatchSource:0}: Error finding container b5597f7e16c010f932748244d694a6afa0a68d54c263e20549f6f52e31ad6b42: Status 404 returned error can't find the container with id b5597f7e16c010f932748244d694a6afa0a68d54c263e20549f6f52e31ad6b42 Mar 17 09:30:51 crc kubenswrapper[4813]: I0317 09:30:51.652371 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:30:51 crc kubenswrapper[4813]: I0317 09:30:51.659798 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7fdb788585-bnbmf"] Mar 17 09:30:51 crc kubenswrapper[4813]: I0317 09:30:51.666537 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-jtbdz"] Mar 17 09:30:51 crc kubenswrapper[4813]: W0317 09:30:51.707072 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c6cd309_922e_4be6_b2d9_12d2aca54b78.slice/crio-d7ee9fb0dab48c1954c52b2160ee734071056c2ba77403f85c1e40679cdd5170 WatchSource:0}: Error finding container d7ee9fb0dab48c1954c52b2160ee734071056c2ba77403f85c1e40679cdd5170: Status 404 returned error can't find the container with id d7ee9fb0dab48c1954c52b2160ee734071056c2ba77403f85c1e40679cdd5170 Mar 17 09:30:51 crc kubenswrapper[4813]: I0317 09:30:51.730034 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-mxghx"] Mar 17 09:30:51 crc kubenswrapper[4813]: I0317 09:30:51.745123 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-8654c79df-tc6h4"] Mar 17 09:30:51 crc kubenswrapper[4813]: W0317 09:30:51.754684 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff5c70e8_f0f9_41ab_8f05_db5a8e708dfb.slice/crio-7f011e0d9c060727b8b2b8d4607dd1cdc553e316e675267af93c8ed24f8e05ad WatchSource:0}: Error finding container 7f011e0d9c060727b8b2b8d4607dd1cdc553e316e675267af93c8ed24f8e05ad: Status 404 returned error can't find the container with id 7f011e0d9c060727b8b2b8d4607dd1cdc553e316e675267af93c8ed24f8e05ad Mar 17 09:30:51 crc kubenswrapper[4813]: I0317 09:30:51.763800 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-27g9l"] Mar 17 09:30:51 crc kubenswrapper[4813]: I0317 09:30:51.943644 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-6wc77"] Mar 17 09:30:51 crc kubenswrapper[4813]: I0317 09:30:51.961239 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-hkmc6"] Mar 17 09:30:52 crc kubenswrapper[4813]: I0317 09:30:52.129575 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 09:30:52 crc kubenswrapper[4813]: W0317 09:30:52.143340 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74a2cdb9_0725_4445_ba2a_4889c8c6a680.slice/crio-72014da49f33457fd6097287448ea60f56f75a0f9bcb65ac4844d75a3ef10be9 WatchSource:0}: Error finding container 72014da49f33457fd6097287448ea60f56f75a0f9bcb65ac4844d75a3ef10be9: Status 404 returned error can't find the container with id 72014da49f33457fd6097287448ea60f56f75a0f9bcb65ac4844d75a3ef10be9 Mar 17 09:30:52 crc kubenswrapper[4813]: I0317 09:30:52.225171 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 09:30:52 crc kubenswrapper[4813]: W0317 09:30:52.231562 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb5357557_3288_455b_8997_231364e39026.slice/crio-843bee686481e0d3afe9634eb0ad767777434e38b90a2da92f395ee93215769f WatchSource:0}: Error finding container 843bee686481e0d3afe9634eb0ad767777434e38b90a2da92f395ee93215769f: Status 404 returned error can't find the container with id 843bee686481e0d3afe9634eb0ad767777434e38b90a2da92f395ee93215769f Mar 17 09:30:52 crc kubenswrapper[4813]: I0317 09:30:52.521775 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d927c167-348f-47f8-9860-b21dac3ca309","Type":"ContainerStarted","Data":"9e18c6e47f25e05f0bc375bef10326d7fd8e3b51c7be154866b1604721bfb3da"} Mar 17 09:30:52 crc kubenswrapper[4813]: I0317 09:30:52.533133 4813 generic.go:334] "Generic (PLEG): container finished" podID="cbe63abb-2664-4824-9f02-001b8ec06edb" containerID="efc04e5a41f72bad94293b25b9242cc59e0fb3772c7456d0d13fe066177af27a" exitCode=0 Mar 17 09:30:52 crc kubenswrapper[4813]: I0317 09:30:52.533250 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" event={"ID":"cbe63abb-2664-4824-9f02-001b8ec06edb","Type":"ContainerDied","Data":"efc04e5a41f72bad94293b25b9242cc59e0fb3772c7456d0d13fe066177af27a"} Mar 17 09:30:52 crc kubenswrapper[4813]: I0317 09:30:52.536761 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-27g9l" event={"ID":"41809604-4a36-4453-b381-66b4c5cf1c43","Type":"ContainerStarted","Data":"4bd5f71c52301a2b0cf8e50675e7fb31caf20086f6fdd0ee80fbd5f75e50fe39"} Mar 17 09:30:52 crc kubenswrapper[4813]: I0317 09:30:52.540405 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jtbdz" event={"ID":"765fce5e-7101-42a5-8ad1-172c967fd2c3","Type":"ContainerStarted","Data":"3e5b3f1123b7a79595271d2af94d410f821a75be59a0cf9ea6e77c1f02f6edee"} Mar 17 09:30:52 crc kubenswrapper[4813]: I0317 09:30:52.540460 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jtbdz" event={"ID":"765fce5e-7101-42a5-8ad1-172c967fd2c3","Type":"ContainerStarted","Data":"9ee3f1bc7abba56e4dc87407f046784eef77995844d29d420038fbcba4e7e951"} Mar 17 09:30:52 crc kubenswrapper[4813]: I0317 09:30:52.543925 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b5357557-3288-455b-8997-231364e39026","Type":"ContainerStarted","Data":"843bee686481e0d3afe9634eb0ad767777434e38b90a2da92f395ee93215769f"} Mar 17 09:30:52 crc kubenswrapper[4813]: I0317 09:30:52.549458 4813 generic.go:334] "Generic (PLEG): container finished" podID="1596452f-363d-4d92-82de-f96f1f83993a" containerID="4fe25e5c2f86364582e41039c2cd8fbf6c091504a790e69bcbb74f5da0ff3642" exitCode=0 Mar 17 09:30:52 crc kubenswrapper[4813]: I0317 09:30:52.549558 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" event={"ID":"1596452f-363d-4d92-82de-f96f1f83993a","Type":"ContainerDied","Data":"4fe25e5c2f86364582e41039c2cd8fbf6c091504a790e69bcbb74f5da0ff3642"} Mar 17 09:30:52 crc kubenswrapper[4813]: I0317 09:30:52.549716 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" event={"ID":"1596452f-363d-4d92-82de-f96f1f83993a","Type":"ContainerStarted","Data":"b393b89dea2a6d9b43f6704b679c1708b47ccc5c4c66126d4a6d9850df1b5086"} Mar 17 09:30:52 crc kubenswrapper[4813]: I0317 09:30:52.554742 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fdb788585-bnbmf" event={"ID":"93e11613-c787-402d-aabb-384aedd72deb","Type":"ContainerStarted","Data":"b5597f7e16c010f932748244d694a6afa0a68d54c263e20549f6f52e31ad6b42"} Mar 17 09:30:52 crc kubenswrapper[4813]: I0317 09:30:52.559902 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-mxghx" event={"ID":"5c6cd309-922e-4be6-b2d9-12d2aca54b78","Type":"ContainerStarted","Data":"d7ee9fb0dab48c1954c52b2160ee734071056c2ba77403f85c1e40679cdd5170"} Mar 17 09:30:52 crc kubenswrapper[4813]: I0317 09:30:52.569740 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"74a2cdb9-0725-4445-ba2a-4889c8c6a680","Type":"ContainerStarted","Data":"72014da49f33457fd6097287448ea60f56f75a0f9bcb65ac4844d75a3ef10be9"} Mar 17 09:30:52 crc kubenswrapper[4813]: I0317 09:30:52.573571 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hkmc6" event={"ID":"9c613a9b-36b2-437d-9c11-252d7d8ce910","Type":"ContainerStarted","Data":"ea9e78643d9a4c01a60114a74a0342ff8064039e5a3274d16ce37f85fec69a95"} Mar 17 09:30:52 crc kubenswrapper[4813]: I0317 09:30:52.576508 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8654c79df-tc6h4" event={"ID":"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb","Type":"ContainerStarted","Data":"7f011e0d9c060727b8b2b8d4607dd1cdc553e316e675267af93c8ed24f8e05ad"} Mar 17 09:30:52 crc kubenswrapper[4813]: I0317 09:30:52.628475 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-jtbdz" podStartSLOduration=3.628461197 podStartE2EDuration="3.628461197s" podCreationTimestamp="2026-03-17 09:30:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:30:52.624789521 +0000 UTC m=+1274.725593020" watchObservedRunningTime="2026-03-17 09:30:52.628461197 +0000 UTC m=+1274.729264696" Mar 17 09:30:52 crc kubenswrapper[4813]: I0317 09:30:52.969064 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:52.994940 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7fdb788585-bnbmf"] Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.120981 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7b8d767f8c-4ql8v"] Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.122548 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b8d767f8c-4ql8v" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.150247 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.163140 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7b8d767f8c-4ql8v"] Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.176843 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.261578 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-horizon-secret-key\") pod \"horizon-7b8d767f8c-4ql8v\" (UID: \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\") " pod="openstack/horizon-7b8d767f8c-4ql8v" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.261908 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-config-data\") pod \"horizon-7b8d767f8c-4ql8v\" (UID: \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\") " pod="openstack/horizon-7b8d767f8c-4ql8v" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.261936 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-logs\") pod \"horizon-7b8d767f8c-4ql8v\" (UID: \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\") " pod="openstack/horizon-7b8d767f8c-4ql8v" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.262170 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-scripts\") pod \"horizon-7b8d767f8c-4ql8v\" (UID: \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\") " pod="openstack/horizon-7b8d767f8c-4ql8v" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.262292 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqbwg\" (UniqueName: \"kubernetes.io/projected/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-kube-api-access-gqbwg\") pod \"horizon-7b8d767f8c-4ql8v\" (UID: \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\") " pod="openstack/horizon-7b8d767f8c-4ql8v" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.365631 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-horizon-secret-key\") pod \"horizon-7b8d767f8c-4ql8v\" (UID: \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\") " pod="openstack/horizon-7b8d767f8c-4ql8v" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.365686 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-config-data\") pod \"horizon-7b8d767f8c-4ql8v\" (UID: \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\") " pod="openstack/horizon-7b8d767f8c-4ql8v" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.365710 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-logs\") pod \"horizon-7b8d767f8c-4ql8v\" (UID: \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\") " pod="openstack/horizon-7b8d767f8c-4ql8v" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.365788 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-scripts\") pod \"horizon-7b8d767f8c-4ql8v\" (UID: \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\") " pod="openstack/horizon-7b8d767f8c-4ql8v" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.365860 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqbwg\" (UniqueName: \"kubernetes.io/projected/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-kube-api-access-gqbwg\") pod \"horizon-7b8d767f8c-4ql8v\" (UID: \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\") " pod="openstack/horizon-7b8d767f8c-4ql8v" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.368974 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-scripts\") pod \"horizon-7b8d767f8c-4ql8v\" (UID: \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\") " pod="openstack/horizon-7b8d767f8c-4ql8v" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.370903 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-config-data\") pod \"horizon-7b8d767f8c-4ql8v\" (UID: \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\") " pod="openstack/horizon-7b8d767f8c-4ql8v" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.370905 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-logs\") pod \"horizon-7b8d767f8c-4ql8v\" (UID: \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\") " pod="openstack/horizon-7b8d767f8c-4ql8v" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.389084 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqbwg\" (UniqueName: \"kubernetes.io/projected/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-kube-api-access-gqbwg\") pod \"horizon-7b8d767f8c-4ql8v\" (UID: \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\") " pod="openstack/horizon-7b8d767f8c-4ql8v" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.391907 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-horizon-secret-key\") pod \"horizon-7b8d767f8c-4ql8v\" (UID: \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\") " pod="openstack/horizon-7b8d767f8c-4ql8v" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.517233 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b8d767f8c-4ql8v" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.588712 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.600138 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" event={"ID":"cbe63abb-2664-4824-9f02-001b8ec06edb","Type":"ContainerDied","Data":"9a257ddace2c4e3501802f0370f9ae08b1e317ec5a3af9863d6206629b0b60c9"} Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.600191 4813 scope.go:117] "RemoveContainer" containerID="efc04e5a41f72bad94293b25b9242cc59e0fb3772c7456d0d13fe066177af27a" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.639509 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" event={"ID":"1596452f-363d-4d92-82de-f96f1f83993a","Type":"ContainerStarted","Data":"77b92a89118e3453f1d95103179cd5750125b99f7124117d4357e999f4f5584b"} Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.639807 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.642901 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b5357557-3288-455b-8997-231364e39026","Type":"ContainerStarted","Data":"5414e35e564e65cca361dd2fbe19228e6d27a7247e876a06d716dd3d48c9cd85"} Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.656994 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" podStartSLOduration=3.656974685 podStartE2EDuration="3.656974685s" podCreationTimestamp="2026-03-17 09:30:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:30:53.655731055 +0000 UTC m=+1275.756534554" watchObservedRunningTime="2026-03-17 09:30:53.656974685 +0000 UTC m=+1275.757778184" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.669488 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-ovsdbserver-sb\") pod \"cbe63abb-2664-4824-9f02-001b8ec06edb\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.669574 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-ovsdbserver-nb\") pod \"cbe63abb-2664-4824-9f02-001b8ec06edb\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.669631 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trs7m\" (UniqueName: \"kubernetes.io/projected/cbe63abb-2664-4824-9f02-001b8ec06edb-kube-api-access-trs7m\") pod \"cbe63abb-2664-4824-9f02-001b8ec06edb\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.669662 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-config\") pod \"cbe63abb-2664-4824-9f02-001b8ec06edb\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.669694 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-dns-swift-storage-0\") pod \"cbe63abb-2664-4824-9f02-001b8ec06edb\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.669709 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-dns-svc\") pod \"cbe63abb-2664-4824-9f02-001b8ec06edb\" (UID: \"cbe63abb-2664-4824-9f02-001b8ec06edb\") " Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.678819 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbe63abb-2664-4824-9f02-001b8ec06edb-kube-api-access-trs7m" (OuterVolumeSpecName: "kube-api-access-trs7m") pod "cbe63abb-2664-4824-9f02-001b8ec06edb" (UID: "cbe63abb-2664-4824-9f02-001b8ec06edb"). InnerVolumeSpecName "kube-api-access-trs7m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.703845 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cbe63abb-2664-4824-9f02-001b8ec06edb" (UID: "cbe63abb-2664-4824-9f02-001b8ec06edb"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.712059 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cbe63abb-2664-4824-9f02-001b8ec06edb" (UID: "cbe63abb-2664-4824-9f02-001b8ec06edb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.724486 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-config" (OuterVolumeSpecName: "config") pod "cbe63abb-2664-4824-9f02-001b8ec06edb" (UID: "cbe63abb-2664-4824-9f02-001b8ec06edb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.725214 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cbe63abb-2664-4824-9f02-001b8ec06edb" (UID: "cbe63abb-2664-4824-9f02-001b8ec06edb"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.729489 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cbe63abb-2664-4824-9f02-001b8ec06edb" (UID: "cbe63abb-2664-4824-9f02-001b8ec06edb"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.771475 4813 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.771502 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.771512 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.771520 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.771530 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trs7m\" (UniqueName: \"kubernetes.io/projected/cbe63abb-2664-4824-9f02-001b8ec06edb-kube-api-access-trs7m\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:53 crc kubenswrapper[4813]: I0317 09:30:53.771538 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbe63abb-2664-4824-9f02-001b8ec06edb-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:54 crc kubenswrapper[4813]: I0317 09:30:54.024302 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7b8d767f8c-4ql8v"] Mar 17 09:30:54 crc kubenswrapper[4813]: I0317 09:30:54.657247 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b5357557-3288-455b-8997-231364e39026","Type":"ContainerStarted","Data":"35f9ebd3635ee5e437646d3b56eb3b63faffa3c8b7a8482e06ba9996616378b2"} Mar 17 09:30:54 crc kubenswrapper[4813]: I0317 09:30:54.657330 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b5357557-3288-455b-8997-231364e39026" containerName="glance-log" containerID="cri-o://5414e35e564e65cca361dd2fbe19228e6d27a7247e876a06d716dd3d48c9cd85" gracePeriod=30 Mar 17 09:30:54 crc kubenswrapper[4813]: I0317 09:30:54.657766 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="b5357557-3288-455b-8997-231364e39026" containerName="glance-httpd" containerID="cri-o://35f9ebd3635ee5e437646d3b56eb3b63faffa3c8b7a8482e06ba9996616378b2" gracePeriod=30 Mar 17 09:30:54 crc kubenswrapper[4813]: I0317 09:30:54.659243 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8d767f8c-4ql8v" event={"ID":"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6","Type":"ContainerStarted","Data":"7f903dc84603d24335df664545e8281ff4a8058a5e39da023b0f8738a752c373"} Mar 17 09:30:54 crc kubenswrapper[4813]: I0317 09:30:54.668822 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"74a2cdb9-0725-4445-ba2a-4889c8c6a680","Type":"ContainerStarted","Data":"78957fb95f50e56eb4f26044efc8ae79a83eba093fa1cf6bdda080dce5462c48"} Mar 17 09:30:54 crc kubenswrapper[4813]: I0317 09:30:54.668876 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"74a2cdb9-0725-4445-ba2a-4889c8c6a680","Type":"ContainerStarted","Data":"ac1372cbb3dc9f6e712bb96188d42cccb67594210dac077aa87ba2c1d65db543"} Mar 17 09:30:54 crc kubenswrapper[4813]: I0317 09:30:54.669031 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="74a2cdb9-0725-4445-ba2a-4889c8c6a680" containerName="glance-log" containerID="cri-o://ac1372cbb3dc9f6e712bb96188d42cccb67594210dac077aa87ba2c1d65db543" gracePeriod=30 Mar 17 09:30:54 crc kubenswrapper[4813]: I0317 09:30:54.669354 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="74a2cdb9-0725-4445-ba2a-4889c8c6a680" containerName="glance-httpd" containerID="cri-o://78957fb95f50e56eb4f26044efc8ae79a83eba093fa1cf6bdda080dce5462c48" gracePeriod=30 Mar 17 09:30:54 crc kubenswrapper[4813]: I0317 09:30:54.682516 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6" Mar 17 09:30:54 crc kubenswrapper[4813]: I0317 09:30:54.722617 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.722584279 podStartE2EDuration="4.722584279s" podCreationTimestamp="2026-03-17 09:30:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:30:54.68286989 +0000 UTC m=+1276.783673389" watchObservedRunningTime="2026-03-17 09:30:54.722584279 +0000 UTC m=+1276.823387778" Mar 17 09:30:54 crc kubenswrapper[4813]: I0317 09:30:54.727007 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.726993987 podStartE2EDuration="4.726993987s" podCreationTimestamp="2026-03-17 09:30:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:30:54.718648075 +0000 UTC m=+1276.819451564" watchObservedRunningTime="2026-03-17 09:30:54.726993987 +0000 UTC m=+1276.827797486" Mar 17 09:30:54 crc kubenswrapper[4813]: I0317 09:30:54.784196 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6"] Mar 17 09:30:54 crc kubenswrapper[4813]: I0317 09:30:54.797293 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c5cc7c5ff-cfnp6"] Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.434207 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.510362 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.510450 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-combined-ca-bundle\") pod \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.510539 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-config-data\") pod \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.510574 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74a2cdb9-0725-4445-ba2a-4889c8c6a680-httpd-run\") pod \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.510771 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74a2cdb9-0725-4445-ba2a-4889c8c6a680-logs\") pod \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.510811 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-internal-tls-certs\") pod \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.510836 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-scripts\") pod \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.510856 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.513373 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74a2cdb9-0725-4445-ba2a-4889c8c6a680-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "74a2cdb9-0725-4445-ba2a-4889c8c6a680" (UID: "74a2cdb9-0725-4445-ba2a-4889c8c6a680"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.530853 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74a2cdb9-0725-4445-ba2a-4889c8c6a680-logs" (OuterVolumeSpecName: "logs") pod "74a2cdb9-0725-4445-ba2a-4889c8c6a680" (UID: "74a2cdb9-0725-4445-ba2a-4889c8c6a680"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.535323 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "74a2cdb9-0725-4445-ba2a-4889c8c6a680" (UID: "74a2cdb9-0725-4445-ba2a-4889c8c6a680"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.537279 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-scripts" (OuterVolumeSpecName: "scripts") pod "74a2cdb9-0725-4445-ba2a-4889c8c6a680" (UID: "74a2cdb9-0725-4445-ba2a-4889c8c6a680"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.613134 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7gc2g\" (UniqueName: \"kubernetes.io/projected/74a2cdb9-0725-4445-ba2a-4889c8c6a680-kube-api-access-7gc2g\") pod \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\" (UID: \"74a2cdb9-0725-4445-ba2a-4889c8c6a680\") " Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.613738 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/74a2cdb9-0725-4445-ba2a-4889c8c6a680-logs\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.613752 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.613770 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.613781 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/74a2cdb9-0725-4445-ba2a-4889c8c6a680-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.643176 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "74a2cdb9-0725-4445-ba2a-4889c8c6a680" (UID: "74a2cdb9-0725-4445-ba2a-4889c8c6a680"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.643725 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "74a2cdb9-0725-4445-ba2a-4889c8c6a680" (UID: "74a2cdb9-0725-4445-ba2a-4889c8c6a680"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.646854 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74a2cdb9-0725-4445-ba2a-4889c8c6a680-kube-api-access-7gc2g" (OuterVolumeSpecName: "kube-api-access-7gc2g") pod "74a2cdb9-0725-4445-ba2a-4889c8c6a680" (UID: "74a2cdb9-0725-4445-ba2a-4889c8c6a680"). InnerVolumeSpecName "kube-api-access-7gc2g". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.670790 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-config-data" (OuterVolumeSpecName: "config-data") pod "74a2cdb9-0725-4445-ba2a-4889c8c6a680" (UID: "74a2cdb9-0725-4445-ba2a-4889c8c6a680"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.683941 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.725708 4813 generic.go:334] "Generic (PLEG): container finished" podID="b5357557-3288-455b-8997-231364e39026" containerID="35f9ebd3635ee5e437646d3b56eb3b63faffa3c8b7a8482e06ba9996616378b2" exitCode=0 Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.725740 4813 generic.go:334] "Generic (PLEG): container finished" podID="b5357557-3288-455b-8997-231364e39026" containerID="5414e35e564e65cca361dd2fbe19228e6d27a7247e876a06d716dd3d48c9cd85" exitCode=143 Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.725784 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b5357557-3288-455b-8997-231364e39026","Type":"ContainerDied","Data":"35f9ebd3635ee5e437646d3b56eb3b63faffa3c8b7a8482e06ba9996616378b2"} Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.725809 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b5357557-3288-455b-8997-231364e39026","Type":"ContainerDied","Data":"5414e35e564e65cca361dd2fbe19228e6d27a7247e876a06d716dd3d48c9cd85"} Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.725820 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"b5357557-3288-455b-8997-231364e39026","Type":"ContainerDied","Data":"843bee686481e0d3afe9634eb0ad767777434e38b90a2da92f395ee93215769f"} Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.725835 4813 scope.go:117] "RemoveContainer" containerID="35f9ebd3635ee5e437646d3b56eb3b63faffa3c8b7a8482e06ba9996616378b2" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.725953 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.731011 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b5357557-3288-455b-8997-231364e39026-httpd-run\") pod \"b5357557-3288-455b-8997-231364e39026\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.731049 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"b5357557-3288-455b-8997-231364e39026\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.731133 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-scripts\") pod \"b5357557-3288-455b-8997-231364e39026\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.731160 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-combined-ca-bundle\") pod \"b5357557-3288-455b-8997-231364e39026\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.731190 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-config-data\") pod \"b5357557-3288-455b-8997-231364e39026\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.731328 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-public-tls-certs\") pod \"b5357557-3288-455b-8997-231364e39026\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.731361 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5357557-3288-455b-8997-231364e39026-logs\") pod \"b5357557-3288-455b-8997-231364e39026\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.731398 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b979q\" (UniqueName: \"kubernetes.io/projected/b5357557-3288-455b-8997-231364e39026-kube-api-access-b979q\") pod \"b5357557-3288-455b-8997-231364e39026\" (UID: \"b5357557-3288-455b-8997-231364e39026\") " Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.731742 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.731759 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7gc2g\" (UniqueName: \"kubernetes.io/projected/74a2cdb9-0725-4445-ba2a-4889c8c6a680-kube-api-access-7gc2g\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.731768 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.731778 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.731786 4813 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/74a2cdb9-0725-4445-ba2a-4889c8c6a680-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.735179 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5357557-3288-455b-8997-231364e39026-logs" (OuterVolumeSpecName: "logs") pod "b5357557-3288-455b-8997-231364e39026" (UID: "b5357557-3288-455b-8997-231364e39026"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.735370 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5357557-3288-455b-8997-231364e39026-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b5357557-3288-455b-8997-231364e39026" (UID: "b5357557-3288-455b-8997-231364e39026"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.735446 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5357557-3288-455b-8997-231364e39026-kube-api-access-b979q" (OuterVolumeSpecName: "kube-api-access-b979q") pod "b5357557-3288-455b-8997-231364e39026" (UID: "b5357557-3288-455b-8997-231364e39026"). InnerVolumeSpecName "kube-api-access-b979q". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.744563 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "b5357557-3288-455b-8997-231364e39026" (UID: "b5357557-3288-455b-8997-231364e39026"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.744797 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-scripts" (OuterVolumeSpecName: "scripts") pod "b5357557-3288-455b-8997-231364e39026" (UID: "b5357557-3288-455b-8997-231364e39026"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.762385 4813 generic.go:334] "Generic (PLEG): container finished" podID="74a2cdb9-0725-4445-ba2a-4889c8c6a680" containerID="78957fb95f50e56eb4f26044efc8ae79a83eba093fa1cf6bdda080dce5462c48" exitCode=143 Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.762617 4813 generic.go:334] "Generic (PLEG): container finished" podID="74a2cdb9-0725-4445-ba2a-4889c8c6a680" containerID="ac1372cbb3dc9f6e712bb96188d42cccb67594210dac077aa87ba2c1d65db543" exitCode=143 Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.762541 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.762511 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"74a2cdb9-0725-4445-ba2a-4889c8c6a680","Type":"ContainerDied","Data":"78957fb95f50e56eb4f26044efc8ae79a83eba093fa1cf6bdda080dce5462c48"} Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.762747 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"74a2cdb9-0725-4445-ba2a-4889c8c6a680","Type":"ContainerDied","Data":"ac1372cbb3dc9f6e712bb96188d42cccb67594210dac077aa87ba2c1d65db543"} Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.762772 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"74a2cdb9-0725-4445-ba2a-4889c8c6a680","Type":"ContainerDied","Data":"72014da49f33457fd6097287448ea60f56f75a0f9bcb65ac4844d75a3ef10be9"} Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.773970 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b5357557-3288-455b-8997-231364e39026" (UID: "b5357557-3288-455b-8997-231364e39026"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.790057 4813 scope.go:117] "RemoveContainer" containerID="5414e35e564e65cca361dd2fbe19228e6d27a7247e876a06d716dd3d48c9cd85" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.805894 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.822747 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.833494 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b5357557-3288-455b-8997-231364e39026-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.833551 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.833562 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.833570 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.833593 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b5357557-3288-455b-8997-231364e39026-logs\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.833615 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b979q\" (UniqueName: \"kubernetes.io/projected/b5357557-3288-455b-8997-231364e39026-kube-api-access-b979q\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.839112 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.840117 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b5357557-3288-455b-8997-231364e39026" (UID: "b5357557-3288-455b-8997-231364e39026"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:55 crc kubenswrapper[4813]: E0317 09:30:55.840837 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbe63abb-2664-4824-9f02-001b8ec06edb" containerName="init" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.840883 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbe63abb-2664-4824-9f02-001b8ec06edb" containerName="init" Mar 17 09:30:55 crc kubenswrapper[4813]: E0317 09:30:55.840904 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a2cdb9-0725-4445-ba2a-4889c8c6a680" containerName="glance-httpd" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.840911 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a2cdb9-0725-4445-ba2a-4889c8c6a680" containerName="glance-httpd" Mar 17 09:30:55 crc kubenswrapper[4813]: E0317 09:30:55.840961 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5357557-3288-455b-8997-231364e39026" containerName="glance-log" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.840969 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5357557-3288-455b-8997-231364e39026" containerName="glance-log" Mar 17 09:30:55 crc kubenswrapper[4813]: E0317 09:30:55.840986 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74a2cdb9-0725-4445-ba2a-4889c8c6a680" containerName="glance-log" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.840991 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="74a2cdb9-0725-4445-ba2a-4889c8c6a680" containerName="glance-log" Mar 17 09:30:55 crc kubenswrapper[4813]: E0317 09:30:55.841007 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5357557-3288-455b-8997-231364e39026" containerName="glance-httpd" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.841014 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5357557-3288-455b-8997-231364e39026" containerName="glance-httpd" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.841244 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbe63abb-2664-4824-9f02-001b8ec06edb" containerName="init" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.841288 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="74a2cdb9-0725-4445-ba2a-4889c8c6a680" containerName="glance-httpd" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.841299 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5357557-3288-455b-8997-231364e39026" containerName="glance-log" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.841309 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5357557-3288-455b-8997-231364e39026" containerName="glance-httpd" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.841320 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="74a2cdb9-0725-4445-ba2a-4889c8c6a680" containerName="glance-log" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.851120 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.851225 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.855988 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.856249 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.860706 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.887786 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-config-data" (OuterVolumeSpecName: "config-data") pod "b5357557-3288-455b-8997-231364e39026" (UID: "b5357557-3288-455b-8997-231364e39026"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.901849 4813 scope.go:117] "RemoveContainer" containerID="35f9ebd3635ee5e437646d3b56eb3b63faffa3c8b7a8482e06ba9996616378b2" Mar 17 09:30:55 crc kubenswrapper[4813]: E0317 09:30:55.903018 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35f9ebd3635ee5e437646d3b56eb3b63faffa3c8b7a8482e06ba9996616378b2\": container with ID starting with 35f9ebd3635ee5e437646d3b56eb3b63faffa3c8b7a8482e06ba9996616378b2 not found: ID does not exist" containerID="35f9ebd3635ee5e437646d3b56eb3b63faffa3c8b7a8482e06ba9996616378b2" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.903058 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35f9ebd3635ee5e437646d3b56eb3b63faffa3c8b7a8482e06ba9996616378b2"} err="failed to get container status \"35f9ebd3635ee5e437646d3b56eb3b63faffa3c8b7a8482e06ba9996616378b2\": rpc error: code = NotFound desc = could not find container \"35f9ebd3635ee5e437646d3b56eb3b63faffa3c8b7a8482e06ba9996616378b2\": container with ID starting with 35f9ebd3635ee5e437646d3b56eb3b63faffa3c8b7a8482e06ba9996616378b2 not found: ID does not exist" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.903078 4813 scope.go:117] "RemoveContainer" containerID="5414e35e564e65cca361dd2fbe19228e6d27a7247e876a06d716dd3d48c9cd85" Mar 17 09:30:55 crc kubenswrapper[4813]: E0317 09:30:55.903636 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5414e35e564e65cca361dd2fbe19228e6d27a7247e876a06d716dd3d48c9cd85\": container with ID starting with 5414e35e564e65cca361dd2fbe19228e6d27a7247e876a06d716dd3d48c9cd85 not found: ID does not exist" containerID="5414e35e564e65cca361dd2fbe19228e6d27a7247e876a06d716dd3d48c9cd85" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.903675 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5414e35e564e65cca361dd2fbe19228e6d27a7247e876a06d716dd3d48c9cd85"} err="failed to get container status \"5414e35e564e65cca361dd2fbe19228e6d27a7247e876a06d716dd3d48c9cd85\": rpc error: code = NotFound desc = could not find container \"5414e35e564e65cca361dd2fbe19228e6d27a7247e876a06d716dd3d48c9cd85\": container with ID starting with 5414e35e564e65cca361dd2fbe19228e6d27a7247e876a06d716dd3d48c9cd85 not found: ID does not exist" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.903694 4813 scope.go:117] "RemoveContainer" containerID="35f9ebd3635ee5e437646d3b56eb3b63faffa3c8b7a8482e06ba9996616378b2" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.904027 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35f9ebd3635ee5e437646d3b56eb3b63faffa3c8b7a8482e06ba9996616378b2"} err="failed to get container status \"35f9ebd3635ee5e437646d3b56eb3b63faffa3c8b7a8482e06ba9996616378b2\": rpc error: code = NotFound desc = could not find container \"35f9ebd3635ee5e437646d3b56eb3b63faffa3c8b7a8482e06ba9996616378b2\": container with ID starting with 35f9ebd3635ee5e437646d3b56eb3b63faffa3c8b7a8482e06ba9996616378b2 not found: ID does not exist" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.904065 4813 scope.go:117] "RemoveContainer" containerID="5414e35e564e65cca361dd2fbe19228e6d27a7247e876a06d716dd3d48c9cd85" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.904527 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5414e35e564e65cca361dd2fbe19228e6d27a7247e876a06d716dd3d48c9cd85"} err="failed to get container status \"5414e35e564e65cca361dd2fbe19228e6d27a7247e876a06d716dd3d48c9cd85\": rpc error: code = NotFound desc = could not find container \"5414e35e564e65cca361dd2fbe19228e6d27a7247e876a06d716dd3d48c9cd85\": container with ID starting with 5414e35e564e65cca361dd2fbe19228e6d27a7247e876a06d716dd3d48c9cd85 not found: ID does not exist" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.904542 4813 scope.go:117] "RemoveContainer" containerID="78957fb95f50e56eb4f26044efc8ae79a83eba093fa1cf6bdda080dce5462c48" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.935097 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.935140 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:55 crc kubenswrapper[4813]: I0317 09:30:55.935156 4813 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b5357557-3288-455b-8997-231364e39026-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.008229 4813 scope.go:117] "RemoveContainer" containerID="ac1372cbb3dc9f6e712bb96188d42cccb67594210dac077aa87ba2c1d65db543" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.036058 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.036110 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-logs\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.036132 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.036169 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.036196 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8spbx\" (UniqueName: \"kubernetes.io/projected/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-kube-api-access-8spbx\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.036243 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.036259 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.036276 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.042981 4813 scope.go:117] "RemoveContainer" containerID="78957fb95f50e56eb4f26044efc8ae79a83eba093fa1cf6bdda080dce5462c48" Mar 17 09:30:56 crc kubenswrapper[4813]: E0317 09:30:56.043528 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78957fb95f50e56eb4f26044efc8ae79a83eba093fa1cf6bdda080dce5462c48\": container with ID starting with 78957fb95f50e56eb4f26044efc8ae79a83eba093fa1cf6bdda080dce5462c48 not found: ID does not exist" containerID="78957fb95f50e56eb4f26044efc8ae79a83eba093fa1cf6bdda080dce5462c48" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.043567 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78957fb95f50e56eb4f26044efc8ae79a83eba093fa1cf6bdda080dce5462c48"} err="failed to get container status \"78957fb95f50e56eb4f26044efc8ae79a83eba093fa1cf6bdda080dce5462c48\": rpc error: code = NotFound desc = could not find container \"78957fb95f50e56eb4f26044efc8ae79a83eba093fa1cf6bdda080dce5462c48\": container with ID starting with 78957fb95f50e56eb4f26044efc8ae79a83eba093fa1cf6bdda080dce5462c48 not found: ID does not exist" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.043609 4813 scope.go:117] "RemoveContainer" containerID="ac1372cbb3dc9f6e712bb96188d42cccb67594210dac077aa87ba2c1d65db543" Mar 17 09:30:56 crc kubenswrapper[4813]: E0317 09:30:56.044152 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac1372cbb3dc9f6e712bb96188d42cccb67594210dac077aa87ba2c1d65db543\": container with ID starting with ac1372cbb3dc9f6e712bb96188d42cccb67594210dac077aa87ba2c1d65db543 not found: ID does not exist" containerID="ac1372cbb3dc9f6e712bb96188d42cccb67594210dac077aa87ba2c1d65db543" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.044237 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac1372cbb3dc9f6e712bb96188d42cccb67594210dac077aa87ba2c1d65db543"} err="failed to get container status \"ac1372cbb3dc9f6e712bb96188d42cccb67594210dac077aa87ba2c1d65db543\": rpc error: code = NotFound desc = could not find container \"ac1372cbb3dc9f6e712bb96188d42cccb67594210dac077aa87ba2c1d65db543\": container with ID starting with ac1372cbb3dc9f6e712bb96188d42cccb67594210dac077aa87ba2c1d65db543 not found: ID does not exist" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.044271 4813 scope.go:117] "RemoveContainer" containerID="78957fb95f50e56eb4f26044efc8ae79a83eba093fa1cf6bdda080dce5462c48" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.045450 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78957fb95f50e56eb4f26044efc8ae79a83eba093fa1cf6bdda080dce5462c48"} err="failed to get container status \"78957fb95f50e56eb4f26044efc8ae79a83eba093fa1cf6bdda080dce5462c48\": rpc error: code = NotFound desc = could not find container \"78957fb95f50e56eb4f26044efc8ae79a83eba093fa1cf6bdda080dce5462c48\": container with ID starting with 78957fb95f50e56eb4f26044efc8ae79a83eba093fa1cf6bdda080dce5462c48 not found: ID does not exist" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.045483 4813 scope.go:117] "RemoveContainer" containerID="ac1372cbb3dc9f6e712bb96188d42cccb67594210dac077aa87ba2c1d65db543" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.045859 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac1372cbb3dc9f6e712bb96188d42cccb67594210dac077aa87ba2c1d65db543"} err="failed to get container status \"ac1372cbb3dc9f6e712bb96188d42cccb67594210dac077aa87ba2c1d65db543\": rpc error: code = NotFound desc = could not find container \"ac1372cbb3dc9f6e712bb96188d42cccb67594210dac077aa87ba2c1d65db543\": container with ID starting with ac1372cbb3dc9f6e712bb96188d42cccb67594210dac077aa87ba2c1d65db543 not found: ID does not exist" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.067280 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.081026 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.103578 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.105062 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.109768 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.109961 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.111062 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.137707 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.137752 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.137779 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.137848 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.137874 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-logs\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.137915 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.137951 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.138046 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8spbx\" (UniqueName: \"kubernetes.io/projected/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-kube-api-access-8spbx\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.139222 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.139789 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-logs\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.140443 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.143354 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.143567 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-config-data\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.144028 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.147449 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-scripts\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.153891 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8spbx\" (UniqueName: \"kubernetes.io/projected/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-kube-api-access-8spbx\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.164658 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.239641 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzhjg\" (UniqueName: \"kubernetes.io/projected/f01d7c15-58db-48db-84d5-e29b1edb2205-kube-api-access-fzhjg\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.239701 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f01d7c15-58db-48db-84d5-e29b1edb2205-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.239717 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.239745 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-config-data\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.239772 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.239802 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f01d7c15-58db-48db-84d5-e29b1edb2205-logs\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.239966 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-scripts\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.240076 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.288845 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.342370 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f01d7c15-58db-48db-84d5-e29b1edb2205-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.342422 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.342477 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-config-data\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.342505 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.342551 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f01d7c15-58db-48db-84d5-e29b1edb2205-logs\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.342580 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-scripts\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.342634 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.342705 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzhjg\" (UniqueName: \"kubernetes.io/projected/f01d7c15-58db-48db-84d5-e29b1edb2205-kube-api-access-fzhjg\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.343466 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f01d7c15-58db-48db-84d5-e29b1edb2205-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.343737 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f01d7c15-58db-48db-84d5-e29b1edb2205-logs\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.343943 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.349697 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-config-data\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.351284 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.351954 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.355591 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-scripts\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.360836 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzhjg\" (UniqueName: \"kubernetes.io/projected/f01d7c15-58db-48db-84d5-e29b1edb2205-kube-api-access-fzhjg\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.375134 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.429413 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.744020 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74a2cdb9-0725-4445-ba2a-4889c8c6a680" path="/var/lib/kubelet/pods/74a2cdb9-0725-4445-ba2a-4889c8c6a680/volumes" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.745052 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5357557-3288-455b-8997-231364e39026" path="/var/lib/kubelet/pods/b5357557-3288-455b-8997-231364e39026/volumes" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.745623 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbe63abb-2664-4824-9f02-001b8ec06edb" path="/var/lib/kubelet/pods/cbe63abb-2664-4824-9f02-001b8ec06edb/volumes" Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.782906 4813 generic.go:334] "Generic (PLEG): container finished" podID="2728d43b-fbec-4573-9f2f-5d29b57e82d6" containerID="0915e99e93df5119ef52f738f246dcf25eb4dd780225e4b1f659ae48abeb232b" exitCode=0 Mar 17 09:30:56 crc kubenswrapper[4813]: I0317 09:30:56.782945 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-p85pk" event={"ID":"2728d43b-fbec-4573-9f2f-5d29b57e82d6","Type":"ContainerDied","Data":"0915e99e93df5119ef52f738f246dcf25eb4dd780225e4b1f659ae48abeb232b"} Mar 17 09:30:58 crc kubenswrapper[4813]: I0317 09:30:58.970202 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.071654 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-p85pk" event={"ID":"2728d43b-fbec-4573-9f2f-5d29b57e82d6","Type":"ContainerDied","Data":"b49e32ad5b9b4a883def4c563e572945b93adac12f9225d90ec113d7790d0fe1"} Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.071695 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b49e32ad5b9b4a883def4c563e572945b93adac12f9225d90ec113d7790d0fe1" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.071910 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-p85pk" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.170376 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-config-data\") pod \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.170455 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z9xkb\" (UniqueName: \"kubernetes.io/projected/2728d43b-fbec-4573-9f2f-5d29b57e82d6-kube-api-access-z9xkb\") pod \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.171029 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-combined-ca-bundle\") pod \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.171117 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-scripts\") pod \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.171188 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-credential-keys\") pod \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.172455 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-fernet-keys\") pod \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\" (UID: \"2728d43b-fbec-4573-9f2f-5d29b57e82d6\") " Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.178002 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2728d43b-fbec-4573-9f2f-5d29b57e82d6-kube-api-access-z9xkb" (OuterVolumeSpecName: "kube-api-access-z9xkb") pod "2728d43b-fbec-4573-9f2f-5d29b57e82d6" (UID: "2728d43b-fbec-4573-9f2f-5d29b57e82d6"). InnerVolumeSpecName "kube-api-access-z9xkb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.191178 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-scripts" (OuterVolumeSpecName: "scripts") pod "2728d43b-fbec-4573-9f2f-5d29b57e82d6" (UID: "2728d43b-fbec-4573-9f2f-5d29b57e82d6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.197389 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "2728d43b-fbec-4573-9f2f-5d29b57e82d6" (UID: "2728d43b-fbec-4573-9f2f-5d29b57e82d6"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.197527 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "2728d43b-fbec-4573-9f2f-5d29b57e82d6" (UID: "2728d43b-fbec-4573-9f2f-5d29b57e82d6"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.215123 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2728d43b-fbec-4573-9f2f-5d29b57e82d6" (UID: "2728d43b-fbec-4573-9f2f-5d29b57e82d6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.234239 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-config-data" (OuterVolumeSpecName: "config-data") pod "2728d43b-fbec-4573-9f2f-5d29b57e82d6" (UID: "2728d43b-fbec-4573-9f2f-5d29b57e82d6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.284385 4813 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.284417 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.284427 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z9xkb\" (UniqueName: \"kubernetes.io/projected/2728d43b-fbec-4573-9f2f-5d29b57e82d6-kube-api-access-z9xkb\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.284438 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.284447 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.284454 4813 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/2728d43b-fbec-4573-9f2f-5d29b57e82d6-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.957391 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-8654c79df-tc6h4"] Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.980723 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-85fc85fdfb-n469t"] Mar 17 09:30:59 crc kubenswrapper[4813]: E0317 09:30:59.981046 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2728d43b-fbec-4573-9f2f-5d29b57e82d6" containerName="keystone-bootstrap" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.981058 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="2728d43b-fbec-4573-9f2f-5d29b57e82d6" containerName="keystone-bootstrap" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.981220 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="2728d43b-fbec-4573-9f2f-5d29b57e82d6" containerName="keystone-bootstrap" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.982047 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.985386 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.996160 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0a52b24b-cc19-4720-a05b-961015362769-horizon-secret-key\") pod \"horizon-85fc85fdfb-n469t\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.996209 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0a52b24b-cc19-4720-a05b-961015362769-config-data\") pod \"horizon-85fc85fdfb-n469t\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.996232 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a52b24b-cc19-4720-a05b-961015362769-logs\") pod \"horizon-85fc85fdfb-n469t\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.996263 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxwcd\" (UniqueName: \"kubernetes.io/projected/0a52b24b-cc19-4720-a05b-961015362769-kube-api-access-fxwcd\") pod \"horizon-85fc85fdfb-n469t\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.996583 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a52b24b-cc19-4720-a05b-961015362769-combined-ca-bundle\") pod \"horizon-85fc85fdfb-n469t\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.996658 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0a52b24b-cc19-4720-a05b-961015362769-scripts\") pod \"horizon-85fc85fdfb-n469t\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:30:59 crc kubenswrapper[4813]: I0317 09:30:59.996716 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a52b24b-cc19-4720-a05b-961015362769-horizon-tls-certs\") pod \"horizon-85fc85fdfb-n469t\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.004211 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.007537 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85fc85fdfb-n469t"] Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.050660 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7b8d767f8c-4ql8v"] Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.090531 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5dc999f6f8-t94dl"] Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.092023 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.096768 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.098554 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0a52b24b-cc19-4720-a05b-961015362769-horizon-secret-key\") pod \"horizon-85fc85fdfb-n469t\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.098650 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0a52b24b-cc19-4720-a05b-961015362769-config-data\") pod \"horizon-85fc85fdfb-n469t\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.098678 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a52b24b-cc19-4720-a05b-961015362769-logs\") pod \"horizon-85fc85fdfb-n469t\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.098765 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxwcd\" (UniqueName: \"kubernetes.io/projected/0a52b24b-cc19-4720-a05b-961015362769-kube-api-access-fxwcd\") pod \"horizon-85fc85fdfb-n469t\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.098897 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a52b24b-cc19-4720-a05b-961015362769-combined-ca-bundle\") pod \"horizon-85fc85fdfb-n469t\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.098969 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0a52b24b-cc19-4720-a05b-961015362769-scripts\") pod \"horizon-85fc85fdfb-n469t\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.099007 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a52b24b-cc19-4720-a05b-961015362769-horizon-tls-certs\") pod \"horizon-85fc85fdfb-n469t\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.099498 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a52b24b-cc19-4720-a05b-961015362769-logs\") pod \"horizon-85fc85fdfb-n469t\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.099759 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0a52b24b-cc19-4720-a05b-961015362769-scripts\") pod \"horizon-85fc85fdfb-n469t\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.099897 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0a52b24b-cc19-4720-a05b-961015362769-config-data\") pod \"horizon-85fc85fdfb-n469t\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.102299 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5dc999f6f8-t94dl"] Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.103571 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a52b24b-cc19-4720-a05b-961015362769-horizon-tls-certs\") pod \"horizon-85fc85fdfb-n469t\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.108816 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a52b24b-cc19-4720-a05b-961015362769-combined-ca-bundle\") pod \"horizon-85fc85fdfb-n469t\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.122225 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0a52b24b-cc19-4720-a05b-961015362769-horizon-secret-key\") pod \"horizon-85fc85fdfb-n469t\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.123036 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxwcd\" (UniqueName: \"kubernetes.io/projected/0a52b24b-cc19-4720-a05b-961015362769-kube-api-access-fxwcd\") pod \"horizon-85fc85fdfb-n469t\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.130737 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-p85pk"] Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.148980 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-p85pk"] Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.189756 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-fdxt8"] Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.191587 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.195999 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.196206 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.196373 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.196435 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-tnvpg" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.196466 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.200106 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf22966e-516d-40c2-975c-c3e41122b8d2-combined-ca-bundle\") pod \"horizon-5dc999f6f8-t94dl\" (UID: \"bf22966e-516d-40c2-975c-c3e41122b8d2\") " pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.200149 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-credential-keys\") pod \"keystone-bootstrap-fdxt8\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.200169 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-combined-ca-bundle\") pod \"keystone-bootstrap-fdxt8\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.200188 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf22966e-516d-40c2-975c-c3e41122b8d2-horizon-tls-certs\") pod \"horizon-5dc999f6f8-t94dl\" (UID: \"bf22966e-516d-40c2-975c-c3e41122b8d2\") " pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.200229 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-fernet-keys\") pod \"keystone-bootstrap-fdxt8\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.200283 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-scripts\") pod \"keystone-bootstrap-fdxt8\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.200332 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf22966e-516d-40c2-975c-c3e41122b8d2-config-data\") pod \"horizon-5dc999f6f8-t94dl\" (UID: \"bf22966e-516d-40c2-975c-c3e41122b8d2\") " pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.200354 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf22966e-516d-40c2-975c-c3e41122b8d2-scripts\") pod \"horizon-5dc999f6f8-t94dl\" (UID: \"bf22966e-516d-40c2-975c-c3e41122b8d2\") " pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.200371 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf22966e-516d-40c2-975c-c3e41122b8d2-logs\") pod \"horizon-5dc999f6f8-t94dl\" (UID: \"bf22966e-516d-40c2-975c-c3e41122b8d2\") " pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.200389 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tl46r\" (UniqueName: \"kubernetes.io/projected/00b72575-1230-444b-9c9b-6c2e2cc921f0-kube-api-access-tl46r\") pod \"keystone-bootstrap-fdxt8\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.200406 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-config-data\") pod \"keystone-bootstrap-fdxt8\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.200434 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87r2g\" (UniqueName: \"kubernetes.io/projected/bf22966e-516d-40c2-975c-c3e41122b8d2-kube-api-access-87r2g\") pod \"horizon-5dc999f6f8-t94dl\" (UID: \"bf22966e-516d-40c2-975c-c3e41122b8d2\") " pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.200464 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bf22966e-516d-40c2-975c-c3e41122b8d2-horizon-secret-key\") pod \"horizon-5dc999f6f8-t94dl\" (UID: \"bf22966e-516d-40c2-975c-c3e41122b8d2\") " pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.208433 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fdxt8"] Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.301863 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf22966e-516d-40c2-975c-c3e41122b8d2-scripts\") pod \"horizon-5dc999f6f8-t94dl\" (UID: \"bf22966e-516d-40c2-975c-c3e41122b8d2\") " pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.301905 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf22966e-516d-40c2-975c-c3e41122b8d2-logs\") pod \"horizon-5dc999f6f8-t94dl\" (UID: \"bf22966e-516d-40c2-975c-c3e41122b8d2\") " pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.301925 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tl46r\" (UniqueName: \"kubernetes.io/projected/00b72575-1230-444b-9c9b-6c2e2cc921f0-kube-api-access-tl46r\") pod \"keystone-bootstrap-fdxt8\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.301956 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-config-data\") pod \"keystone-bootstrap-fdxt8\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.301990 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87r2g\" (UniqueName: \"kubernetes.io/projected/bf22966e-516d-40c2-975c-c3e41122b8d2-kube-api-access-87r2g\") pod \"horizon-5dc999f6f8-t94dl\" (UID: \"bf22966e-516d-40c2-975c-c3e41122b8d2\") " pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.302024 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bf22966e-516d-40c2-975c-c3e41122b8d2-horizon-secret-key\") pod \"horizon-5dc999f6f8-t94dl\" (UID: \"bf22966e-516d-40c2-975c-c3e41122b8d2\") " pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.302048 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf22966e-516d-40c2-975c-c3e41122b8d2-combined-ca-bundle\") pod \"horizon-5dc999f6f8-t94dl\" (UID: \"bf22966e-516d-40c2-975c-c3e41122b8d2\") " pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.302072 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-credential-keys\") pod \"keystone-bootstrap-fdxt8\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.302088 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-combined-ca-bundle\") pod \"keystone-bootstrap-fdxt8\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.302105 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf22966e-516d-40c2-975c-c3e41122b8d2-horizon-tls-certs\") pod \"horizon-5dc999f6f8-t94dl\" (UID: \"bf22966e-516d-40c2-975c-c3e41122b8d2\") " pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.302143 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-fernet-keys\") pod \"keystone-bootstrap-fdxt8\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.302200 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-scripts\") pod \"keystone-bootstrap-fdxt8\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.302230 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf22966e-516d-40c2-975c-c3e41122b8d2-config-data\") pod \"horizon-5dc999f6f8-t94dl\" (UID: \"bf22966e-516d-40c2-975c-c3e41122b8d2\") " pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.302345 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bf22966e-516d-40c2-975c-c3e41122b8d2-logs\") pod \"horizon-5dc999f6f8-t94dl\" (UID: \"bf22966e-516d-40c2-975c-c3e41122b8d2\") " pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.304657 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/bf22966e-516d-40c2-975c-c3e41122b8d2-scripts\") pod \"horizon-5dc999f6f8-t94dl\" (UID: \"bf22966e-516d-40c2-975c-c3e41122b8d2\") " pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.305105 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.308870 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/bf22966e-516d-40c2-975c-c3e41122b8d2-config-data\") pod \"horizon-5dc999f6f8-t94dl\" (UID: \"bf22966e-516d-40c2-975c-c3e41122b8d2\") " pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.310742 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/bf22966e-516d-40c2-975c-c3e41122b8d2-horizon-secret-key\") pod \"horizon-5dc999f6f8-t94dl\" (UID: \"bf22966e-516d-40c2-975c-c3e41122b8d2\") " pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.311499 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-config-data\") pod \"keystone-bootstrap-fdxt8\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.313237 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-fernet-keys\") pod \"keystone-bootstrap-fdxt8\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.317423 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tl46r\" (UniqueName: \"kubernetes.io/projected/00b72575-1230-444b-9c9b-6c2e2cc921f0-kube-api-access-tl46r\") pod \"keystone-bootstrap-fdxt8\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.318779 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/bf22966e-516d-40c2-975c-c3e41122b8d2-horizon-tls-certs\") pod \"horizon-5dc999f6f8-t94dl\" (UID: \"bf22966e-516d-40c2-975c-c3e41122b8d2\") " pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.319896 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-combined-ca-bundle\") pod \"keystone-bootstrap-fdxt8\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.320162 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87r2g\" (UniqueName: \"kubernetes.io/projected/bf22966e-516d-40c2-975c-c3e41122b8d2-kube-api-access-87r2g\") pod \"horizon-5dc999f6f8-t94dl\" (UID: \"bf22966e-516d-40c2-975c-c3e41122b8d2\") " pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.326123 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-scripts\") pod \"keystone-bootstrap-fdxt8\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.333719 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bf22966e-516d-40c2-975c-c3e41122b8d2-combined-ca-bundle\") pod \"horizon-5dc999f6f8-t94dl\" (UID: \"bf22966e-516d-40c2-975c-c3e41122b8d2\") " pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.340358 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-credential-keys\") pod \"keystone-bootstrap-fdxt8\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.490353 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.520119 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.748174 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2728d43b-fbec-4573-9f2f-5d29b57e82d6" path="/var/lib/kubelet/pods/2728d43b-fbec-4573-9f2f-5d29b57e82d6/volumes" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.809783 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.906845 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-crbwc"] Mar 17 09:31:00 crc kubenswrapper[4813]: I0317 09:31:00.907298 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" podUID="549ff45c-ad4e-4456-bf5a-eec6852665b9" containerName="dnsmasq-dns" containerID="cri-o://af39a9ae05bc43d8ae3e27a24eea19bff217feed6796341341ae24991a1c1dc5" gracePeriod=10 Mar 17 09:31:01 crc kubenswrapper[4813]: I0317 09:31:01.112417 4813 generic.go:334] "Generic (PLEG): container finished" podID="549ff45c-ad4e-4456-bf5a-eec6852665b9" containerID="af39a9ae05bc43d8ae3e27a24eea19bff217feed6796341341ae24991a1c1dc5" exitCode=0 Mar 17 09:31:01 crc kubenswrapper[4813]: I0317 09:31:01.112467 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" event={"ID":"549ff45c-ad4e-4456-bf5a-eec6852665b9","Type":"ContainerDied","Data":"af39a9ae05bc43d8ae3e27a24eea19bff217feed6796341341ae24991a1c1dc5"} Mar 17 09:31:02 crc kubenswrapper[4813]: I0317 09:31:02.912847 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" podUID="549ff45c-ad4e-4456-bf5a-eec6852665b9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: connect: connection refused" Mar 17 09:31:07 crc kubenswrapper[4813]: I0317 09:31:07.912440 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" podUID="549ff45c-ad4e-4456-bf5a-eec6852665b9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: connect: connection refused" Mar 17 09:31:10 crc kubenswrapper[4813]: E0317 09:31:10.360551 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Mar 17 09:31:10 crc kubenswrapper[4813]: E0317 09:31:10.361046 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n56dh5bdh6h57h6dhddh56h59ch696h5cbh556h55h5b5h9h7ch67h557h5bdh58fh675h7bh7fh655h9h54bhcch5fdhfch56dh59ch685h94q,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-9h786,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-8654c79df-tc6h4_openstack(ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 09:31:10 crc kubenswrapper[4813]: E0317 09:31:10.363978 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-8654c79df-tc6h4" podUID="ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb" Mar 17 09:31:10 crc kubenswrapper[4813]: E0317 09:31:10.382486 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon:current-podified" Mar 17 09:31:10 crc kubenswrapper[4813]: E0317 09:31:10.382719 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n646h5f6h684hbbh6chfdh8ch5f7h546h5f4h574h585h55fh5dch5bfh588h66hf6h5f5hfdh576h5b5h655h67bhd5h5d7h7ch579h55fh59h57fh58bq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qbx5x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-7fdb788585-bnbmf_openstack(93e11613-c787-402d-aabb-384aedd72deb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 09:31:10 crc kubenswrapper[4813]: E0317 09:31:10.385763 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon:current-podified\\\"\"]" pod="openstack/horizon-7fdb788585-bnbmf" podUID="93e11613-c787-402d-aabb-384aedd72deb" Mar 17 09:31:11 crc kubenswrapper[4813]: I0317 09:31:11.191940 4813 generic.go:334] "Generic (PLEG): container finished" podID="765fce5e-7101-42a5-8ad1-172c967fd2c3" containerID="3e5b3f1123b7a79595271d2af94d410f821a75be59a0cf9ea6e77c1f02f6edee" exitCode=0 Mar 17 09:31:11 crc kubenswrapper[4813]: I0317 09:31:11.192063 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jtbdz" event={"ID":"765fce5e-7101-42a5-8ad1-172c967fd2c3","Type":"ContainerDied","Data":"3e5b3f1123b7a79595271d2af94d410f821a75be59a0cf9ea6e77c1f02f6edee"} Mar 17 09:31:17 crc kubenswrapper[4813]: I0317 09:31:17.912489 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" podUID="549ff45c-ad4e-4456-bf5a-eec6852665b9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: i/o timeout" Mar 17 09:31:17 crc kubenswrapper[4813]: I0317 09:31:17.913501 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:31:19 crc kubenswrapper[4813]: E0317 09:31:19.077615 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified" Mar 17 09:31:19 crc kubenswrapper[4813]: E0317 09:31:19.078048 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lgwjb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-mxghx_openstack(5c6cd309-922e-4be6-b2d9-12d2aca54b78): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 09:31:19 crc kubenswrapper[4813]: E0317 09:31:19.079989 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-mxghx" podUID="5c6cd309-922e-4be6-b2d9-12d2aca54b78" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.244026 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.276028 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jtbdz" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.277738 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fdb788585-bnbmf" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.282184 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-8654c79df-tc6h4" event={"ID":"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb","Type":"ContainerDied","Data":"7f011e0d9c060727b8b2b8d4607dd1cdc553e316e675267af93c8ed24f8e05ad"} Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.282222 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f011e0d9c060727b8b2b8d4607dd1cdc553e316e675267af93c8ed24f8e05ad" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.283468 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8654c79df-tc6h4" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.284188 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-jtbdz" event={"ID":"765fce5e-7101-42a5-8ad1-172c967fd2c3","Type":"ContainerDied","Data":"9ee3f1bc7abba56e4dc87407f046784eef77995844d29d420038fbcba4e7e951"} Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.284222 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ee3f1bc7abba56e4dc87407f046784eef77995844d29d420038fbcba4e7e951" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.284312 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-jtbdz" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.287495 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.287873 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" event={"ID":"549ff45c-ad4e-4456-bf5a-eec6852665b9","Type":"ContainerDied","Data":"7d3434ed12c9880d19cace558c9be3761861a81d2d03cc573a1c7fa04ceba72b"} Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.287921 4813 scope.go:117] "RemoveContainer" containerID="af39a9ae05bc43d8ae3e27a24eea19bff217feed6796341341ae24991a1c1dc5" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.289212 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7fdb788585-bnbmf" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.289258 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7fdb788585-bnbmf" event={"ID":"93e11613-c787-402d-aabb-384aedd72deb","Type":"ContainerDied","Data":"b5597f7e16c010f932748244d694a6afa0a68d54c263e20549f6f52e31ad6b42"} Mar 17 09:31:19 crc kubenswrapper[4813]: E0317 09:31:19.290696 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified\\\"\"" pod="openstack/barbican-db-sync-mxghx" podUID="5c6cd309-922e-4be6-b2d9-12d2aca54b78" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.383050 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbx5x\" (UniqueName: \"kubernetes.io/projected/93e11613-c787-402d-aabb-384aedd72deb-kube-api-access-qbx5x\") pod \"93e11613-c787-402d-aabb-384aedd72deb\" (UID: \"93e11613-c787-402d-aabb-384aedd72deb\") " Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.383102 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-ovsdbserver-sb\") pod \"549ff45c-ad4e-4456-bf5a-eec6852665b9\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.383157 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-dns-swift-storage-0\") pod \"549ff45c-ad4e-4456-bf5a-eec6852665b9\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.383207 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/93e11613-c787-402d-aabb-384aedd72deb-scripts\") pod \"93e11613-c787-402d-aabb-384aedd72deb\" (UID: \"93e11613-c787-402d-aabb-384aedd72deb\") " Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.383223 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/765fce5e-7101-42a5-8ad1-172c967fd2c3-config\") pod \"765fce5e-7101-42a5-8ad1-172c967fd2c3\" (UID: \"765fce5e-7101-42a5-8ad1-172c967fd2c3\") " Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.383289 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-ovsdbserver-nb\") pod \"549ff45c-ad4e-4456-bf5a-eec6852665b9\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.383314 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-dns-svc\") pod \"549ff45c-ad4e-4456-bf5a-eec6852665b9\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.383337 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7q2p2\" (UniqueName: \"kubernetes.io/projected/549ff45c-ad4e-4456-bf5a-eec6852665b9-kube-api-access-7q2p2\") pod \"549ff45c-ad4e-4456-bf5a-eec6852665b9\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.383357 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-config\") pod \"549ff45c-ad4e-4456-bf5a-eec6852665b9\" (UID: \"549ff45c-ad4e-4456-bf5a-eec6852665b9\") " Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.383380 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93e11613-c787-402d-aabb-384aedd72deb-logs\") pod \"93e11613-c787-402d-aabb-384aedd72deb\" (UID: \"93e11613-c787-402d-aabb-384aedd72deb\") " Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.383400 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/765fce5e-7101-42a5-8ad1-172c967fd2c3-combined-ca-bundle\") pod \"765fce5e-7101-42a5-8ad1-172c967fd2c3\" (UID: \"765fce5e-7101-42a5-8ad1-172c967fd2c3\") " Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.383448 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5q4bp\" (UniqueName: \"kubernetes.io/projected/765fce5e-7101-42a5-8ad1-172c967fd2c3-kube-api-access-5q4bp\") pod \"765fce5e-7101-42a5-8ad1-172c967fd2c3\" (UID: \"765fce5e-7101-42a5-8ad1-172c967fd2c3\") " Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.383486 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93e11613-c787-402d-aabb-384aedd72deb-config-data\") pod \"93e11613-c787-402d-aabb-384aedd72deb\" (UID: \"93e11613-c787-402d-aabb-384aedd72deb\") " Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.383516 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/93e11613-c787-402d-aabb-384aedd72deb-horizon-secret-key\") pod \"93e11613-c787-402d-aabb-384aedd72deb\" (UID: \"93e11613-c787-402d-aabb-384aedd72deb\") " Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.385777 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93e11613-c787-402d-aabb-384aedd72deb-scripts" (OuterVolumeSpecName: "scripts") pod "93e11613-c787-402d-aabb-384aedd72deb" (UID: "93e11613-c787-402d-aabb-384aedd72deb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.385812 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/93e11613-c787-402d-aabb-384aedd72deb-logs" (OuterVolumeSpecName: "logs") pod "93e11613-c787-402d-aabb-384aedd72deb" (UID: "93e11613-c787-402d-aabb-384aedd72deb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.386467 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/93e11613-c787-402d-aabb-384aedd72deb-config-data" (OuterVolumeSpecName: "config-data") pod "93e11613-c787-402d-aabb-384aedd72deb" (UID: "93e11613-c787-402d-aabb-384aedd72deb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.389926 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93e11613-c787-402d-aabb-384aedd72deb-kube-api-access-qbx5x" (OuterVolumeSpecName: "kube-api-access-qbx5x") pod "93e11613-c787-402d-aabb-384aedd72deb" (UID: "93e11613-c787-402d-aabb-384aedd72deb"). InnerVolumeSpecName "kube-api-access-qbx5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.392618 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/549ff45c-ad4e-4456-bf5a-eec6852665b9-kube-api-access-7q2p2" (OuterVolumeSpecName: "kube-api-access-7q2p2") pod "549ff45c-ad4e-4456-bf5a-eec6852665b9" (UID: "549ff45c-ad4e-4456-bf5a-eec6852665b9"). InnerVolumeSpecName "kube-api-access-7q2p2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.393776 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/765fce5e-7101-42a5-8ad1-172c967fd2c3-kube-api-access-5q4bp" (OuterVolumeSpecName: "kube-api-access-5q4bp") pod "765fce5e-7101-42a5-8ad1-172c967fd2c3" (UID: "765fce5e-7101-42a5-8ad1-172c967fd2c3"). InnerVolumeSpecName "kube-api-access-5q4bp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.399704 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93e11613-c787-402d-aabb-384aedd72deb-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "93e11613-c787-402d-aabb-384aedd72deb" (UID: "93e11613-c787-402d-aabb-384aedd72deb"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.412835 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/765fce5e-7101-42a5-8ad1-172c967fd2c3-config" (OuterVolumeSpecName: "config") pod "765fce5e-7101-42a5-8ad1-172c967fd2c3" (UID: "765fce5e-7101-42a5-8ad1-172c967fd2c3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.417182 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/765fce5e-7101-42a5-8ad1-172c967fd2c3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "765fce5e-7101-42a5-8ad1-172c967fd2c3" (UID: "765fce5e-7101-42a5-8ad1-172c967fd2c3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.433704 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "549ff45c-ad4e-4456-bf5a-eec6852665b9" (UID: "549ff45c-ad4e-4456-bf5a-eec6852665b9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.439645 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "549ff45c-ad4e-4456-bf5a-eec6852665b9" (UID: "549ff45c-ad4e-4456-bf5a-eec6852665b9"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.440971 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "549ff45c-ad4e-4456-bf5a-eec6852665b9" (UID: "549ff45c-ad4e-4456-bf5a-eec6852665b9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.448125 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-config" (OuterVolumeSpecName: "config") pod "549ff45c-ad4e-4456-bf5a-eec6852665b9" (UID: "549ff45c-ad4e-4456-bf5a-eec6852665b9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.448745 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "549ff45c-ad4e-4456-bf5a-eec6852665b9" (UID: "549ff45c-ad4e-4456-bf5a-eec6852665b9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.484862 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-horizon-secret-key\") pod \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\" (UID: \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\") " Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.484932 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-config-data\") pod \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\" (UID: \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\") " Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.484996 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9h786\" (UniqueName: \"kubernetes.io/projected/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-kube-api-access-9h786\") pod \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\" (UID: \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\") " Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.485026 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-scripts\") pod \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\" (UID: \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\") " Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.485057 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-logs\") pod \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\" (UID: \"ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb\") " Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.485656 4813 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/93e11613-c787-402d-aabb-384aedd72deb-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.485674 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbx5x\" (UniqueName: \"kubernetes.io/projected/93e11613-c787-402d-aabb-384aedd72deb-kube-api-access-qbx5x\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.485687 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.485698 4813 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.485709 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/765fce5e-7101-42a5-8ad1-172c967fd2c3-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.485721 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/93e11613-c787-402d-aabb-384aedd72deb-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.485732 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.485744 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.485757 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7q2p2\" (UniqueName: \"kubernetes.io/projected/549ff45c-ad4e-4456-bf5a-eec6852665b9-kube-api-access-7q2p2\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.485768 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/549ff45c-ad4e-4456-bf5a-eec6852665b9-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.485779 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/93e11613-c787-402d-aabb-384aedd72deb-logs\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.485791 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/765fce5e-7101-42a5-8ad1-172c967fd2c3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.485801 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5q4bp\" (UniqueName: \"kubernetes.io/projected/765fce5e-7101-42a5-8ad1-172c967fd2c3-kube-api-access-5q4bp\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.485811 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/93e11613-c787-402d-aabb-384aedd72deb-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.486655 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-scripts" (OuterVolumeSpecName: "scripts") pod "ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb" (UID: "ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.487380 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-logs" (OuterVolumeSpecName: "logs") pod "ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb" (UID: "ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.488071 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-config-data" (OuterVolumeSpecName: "config-data") pod "ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb" (UID: "ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.488568 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-kube-api-access-9h786" (OuterVolumeSpecName: "kube-api-access-9h786") pod "ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb" (UID: "ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb"). InnerVolumeSpecName "kube-api-access-9h786". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.489052 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb" (UID: "ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.587948 4813 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.587975 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.587986 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9h786\" (UniqueName: \"kubernetes.io/projected/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-kube-api-access-9h786\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.588002 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.588012 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb-logs\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.639682 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-crbwc"] Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.650083 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7ff5475cc9-crbwc"] Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.684943 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7fdb788585-bnbmf"] Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.692156 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7fdb788585-bnbmf"] Mar 17 09:31:19 crc kubenswrapper[4813]: I0317 09:31:19.698082 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.300909 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-8654c79df-tc6h4" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.389569 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-8654c79df-tc6h4"] Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.392418 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-8654c79df-tc6h4"] Mar 17 09:31:20 crc kubenswrapper[4813]: W0317 09:31:20.407275 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf01d7c15_58db_48db_84d5_e29b1edb2205.slice/crio-d925be3c7ce3ffb058eb969923deeb0f10c26ec78625dda621874a5dc90343e4 WatchSource:0}: Error finding container d925be3c7ce3ffb058eb969923deeb0f10c26ec78625dda621874a5dc90343e4: Status 404 returned error can't find the container with id d925be3c7ce3ffb058eb969923deeb0f10c26ec78625dda621874a5dc90343e4 Mar 17 09:31:20 crc kubenswrapper[4813]: E0317 09:31:20.407417 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Mar 17 09:31:20 crc kubenswrapper[4813]: E0317 09:31:20.407548 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-zfpnh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-27g9l_openstack(41809604-4a36-4453-b381-66b4c5cf1c43): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 09:31:20 crc kubenswrapper[4813]: E0317 09:31:20.408725 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-27g9l" podUID="41809604-4a36-4453-b381-66b4c5cf1c43" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.444724 4813 scope.go:117] "RemoveContainer" containerID="d30002deb5bc9b97a656826cac0644cd9994c0466037948281bd22e53aa0dcd2" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.532330 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-mfck7"] Mar 17 09:31:20 crc kubenswrapper[4813]: E0317 09:31:20.533570 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="765fce5e-7101-42a5-8ad1-172c967fd2c3" containerName="neutron-db-sync" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.533591 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="765fce5e-7101-42a5-8ad1-172c967fd2c3" containerName="neutron-db-sync" Mar 17 09:31:20 crc kubenswrapper[4813]: E0317 09:31:20.533614 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="549ff45c-ad4e-4456-bf5a-eec6852665b9" containerName="init" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.533620 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="549ff45c-ad4e-4456-bf5a-eec6852665b9" containerName="init" Mar 17 09:31:20 crc kubenswrapper[4813]: E0317 09:31:20.533629 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="549ff45c-ad4e-4456-bf5a-eec6852665b9" containerName="dnsmasq-dns" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.533635 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="549ff45c-ad4e-4456-bf5a-eec6852665b9" containerName="dnsmasq-dns" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.533788 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="549ff45c-ad4e-4456-bf5a-eec6852665b9" containerName="dnsmasq-dns" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.533813 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="765fce5e-7101-42a5-8ad1-172c967fd2c3" containerName="neutron-db-sync" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.536778 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.551587 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-mfck7"] Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.610575 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-mfck7\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.610670 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r62cj\" (UniqueName: \"kubernetes.io/projected/de2cccf5-7797-4baf-bf2f-d6405959799f-kube-api-access-r62cj\") pod \"dnsmasq-dns-84b966f6c9-mfck7\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.610719 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-mfck7\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.610735 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-mfck7\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.610759 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-mfck7\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.610782 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-config\") pod \"dnsmasq-dns-84b966f6c9-mfck7\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.656551 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-b5495b5bd-g77kp"] Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.663312 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b5495b5bd-g77kp" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.665836 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.665911 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.672128 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-js8px" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.672447 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.717136 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-mfck7\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.717206 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r62cj\" (UniqueName: \"kubernetes.io/projected/de2cccf5-7797-4baf-bf2f-d6405959799f-kube-api-access-r62cj\") pod \"dnsmasq-dns-84b966f6c9-mfck7\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.717254 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-httpd-config\") pod \"neutron-b5495b5bd-g77kp\" (UID: \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\") " pod="openstack/neutron-b5495b5bd-g77kp" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.717283 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-config\") pod \"neutron-b5495b5bd-g77kp\" (UID: \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\") " pod="openstack/neutron-b5495b5bd-g77kp" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.717312 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-ovndb-tls-certs\") pod \"neutron-b5495b5bd-g77kp\" (UID: \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\") " pod="openstack/neutron-b5495b5bd-g77kp" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.717331 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-mfck7\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.717351 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-mfck7\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.717374 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-mfck7\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.717400 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-config\") pod \"dnsmasq-dns-84b966f6c9-mfck7\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.717443 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9fxk\" (UniqueName: \"kubernetes.io/projected/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-kube-api-access-v9fxk\") pod \"neutron-b5495b5bd-g77kp\" (UID: \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\") " pod="openstack/neutron-b5495b5bd-g77kp" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.717475 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-combined-ca-bundle\") pod \"neutron-b5495b5bd-g77kp\" (UID: \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\") " pod="openstack/neutron-b5495b5bd-g77kp" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.718036 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b5495b5bd-g77kp"] Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.718817 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-dns-swift-storage-0\") pod \"dnsmasq-dns-84b966f6c9-mfck7\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.718833 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-ovsdbserver-nb\") pod \"dnsmasq-dns-84b966f6c9-mfck7\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.719047 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-dns-svc\") pod \"dnsmasq-dns-84b966f6c9-mfck7\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.719219 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-config\") pod \"dnsmasq-dns-84b966f6c9-mfck7\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.719332 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-ovsdbserver-sb\") pod \"dnsmasq-dns-84b966f6c9-mfck7\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.745412 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r62cj\" (UniqueName: \"kubernetes.io/projected/de2cccf5-7797-4baf-bf2f-d6405959799f-kube-api-access-r62cj\") pod \"dnsmasq-dns-84b966f6c9-mfck7\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.748655 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="549ff45c-ad4e-4456-bf5a-eec6852665b9" path="/var/lib/kubelet/pods/549ff45c-ad4e-4456-bf5a-eec6852665b9/volumes" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.749870 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93e11613-c787-402d-aabb-384aedd72deb" path="/var/lib/kubelet/pods/93e11613-c787-402d-aabb-384aedd72deb/volumes" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.750357 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb" path="/var/lib/kubelet/pods/ff5c70e8-f0f9-41ab-8f05-db5a8e708dfb/volumes" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.819994 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9fxk\" (UniqueName: \"kubernetes.io/projected/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-kube-api-access-v9fxk\") pod \"neutron-b5495b5bd-g77kp\" (UID: \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\") " pod="openstack/neutron-b5495b5bd-g77kp" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.820455 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-combined-ca-bundle\") pod \"neutron-b5495b5bd-g77kp\" (UID: \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\") " pod="openstack/neutron-b5495b5bd-g77kp" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.820535 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-httpd-config\") pod \"neutron-b5495b5bd-g77kp\" (UID: \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\") " pod="openstack/neutron-b5495b5bd-g77kp" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.820568 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-config\") pod \"neutron-b5495b5bd-g77kp\" (UID: \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\") " pod="openstack/neutron-b5495b5bd-g77kp" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.820650 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-ovndb-tls-certs\") pod \"neutron-b5495b5bd-g77kp\" (UID: \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\") " pod="openstack/neutron-b5495b5bd-g77kp" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.827251 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-ovndb-tls-certs\") pod \"neutron-b5495b5bd-g77kp\" (UID: \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\") " pod="openstack/neutron-b5495b5bd-g77kp" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.835389 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-config\") pod \"neutron-b5495b5bd-g77kp\" (UID: \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\") " pod="openstack/neutron-b5495b5bd-g77kp" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.838094 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-combined-ca-bundle\") pod \"neutron-b5495b5bd-g77kp\" (UID: \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\") " pod="openstack/neutron-b5495b5bd-g77kp" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.838626 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-httpd-config\") pod \"neutron-b5495b5bd-g77kp\" (UID: \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\") " pod="openstack/neutron-b5495b5bd-g77kp" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.842741 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9fxk\" (UniqueName: \"kubernetes.io/projected/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-kube-api-access-v9fxk\") pod \"neutron-b5495b5bd-g77kp\" (UID: \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\") " pod="openstack/neutron-b5495b5bd-g77kp" Mar 17 09:31:20 crc kubenswrapper[4813]: I0317 09:31:20.888490 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:21 crc kubenswrapper[4813]: I0317 09:31:21.002832 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b5495b5bd-g77kp" Mar 17 09:31:21 crc kubenswrapper[4813]: I0317 09:31:21.067588 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-fdxt8"] Mar 17 09:31:21 crc kubenswrapper[4813]: I0317 09:31:21.159121 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 09:31:21 crc kubenswrapper[4813]: I0317 09:31:21.174592 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-85fc85fdfb-n469t"] Mar 17 09:31:21 crc kubenswrapper[4813]: I0317 09:31:21.182578 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5dc999f6f8-t94dl"] Mar 17 09:31:21 crc kubenswrapper[4813]: I0317 09:31:21.323032 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fdxt8" event={"ID":"00b72575-1230-444b-9c9b-6c2e2cc921f0","Type":"ContainerStarted","Data":"95d4fd8e2925edcb3f6c488c377d44212e999965266b1e21641072d605c009c9"} Mar 17 09:31:21 crc kubenswrapper[4813]: I0317 09:31:21.340426 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"563f058b-5cf0-4f1f-91fb-c53c6088a9dc","Type":"ContainerStarted","Data":"6e0e83bd78da580cc4b8d712b6069a9d023a31c09866a1bbbcabd82a6d042944"} Mar 17 09:31:21 crc kubenswrapper[4813]: I0317 09:31:21.344422 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hkmc6" event={"ID":"9c613a9b-36b2-437d-9c11-252d7d8ce910","Type":"ContainerStarted","Data":"3933cb26fc972a9b167b78a49896693ed6615d62ac3e9962d795dde203e9cd3f"} Mar 17 09:31:21 crc kubenswrapper[4813]: I0317 09:31:21.352197 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f01d7c15-58db-48db-84d5-e29b1edb2205","Type":"ContainerStarted","Data":"d925be3c7ce3ffb058eb969923deeb0f10c26ec78625dda621874a5dc90343e4"} Mar 17 09:31:21 crc kubenswrapper[4813]: I0317 09:31:21.358974 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-hkmc6" podStartSLOduration=4.210603306 podStartE2EDuration="31.358958684s" podCreationTimestamp="2026-03-17 09:30:50 +0000 UTC" firstStartedPulling="2026-03-17 09:30:51.969633646 +0000 UTC m=+1274.070437155" lastFinishedPulling="2026-03-17 09:31:19.117989014 +0000 UTC m=+1301.218792533" observedRunningTime="2026-03-17 09:31:21.357504028 +0000 UTC m=+1303.458307527" watchObservedRunningTime="2026-03-17 09:31:21.358958684 +0000 UTC m=+1303.459762183" Mar 17 09:31:21 crc kubenswrapper[4813]: I0317 09:31:21.360991 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8d767f8c-4ql8v" event={"ID":"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6","Type":"ContainerStarted","Data":"c2963271c70c0c8b52d4a2447d8116c4b8f85c2c4475cacbe8918e346d55d5a4"} Mar 17 09:31:21 crc kubenswrapper[4813]: I0317 09:31:21.361034 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8d767f8c-4ql8v" event={"ID":"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6","Type":"ContainerStarted","Data":"d02f3c287e14f9b1f7666d914bf5a8c6fbd8dafffadba75dc06c5be99f1b9a19"} Mar 17 09:31:21 crc kubenswrapper[4813]: I0317 09:31:21.361160 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7b8d767f8c-4ql8v" podUID="8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6" containerName="horizon-log" containerID="cri-o://d02f3c287e14f9b1f7666d914bf5a8c6fbd8dafffadba75dc06c5be99f1b9a19" gracePeriod=30 Mar 17 09:31:21 crc kubenswrapper[4813]: I0317 09:31:21.362839 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7b8d767f8c-4ql8v" podUID="8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6" containerName="horizon" containerID="cri-o://c2963271c70c0c8b52d4a2447d8116c4b8f85c2c4475cacbe8918e346d55d5a4" gracePeriod=30 Mar 17 09:31:21 crc kubenswrapper[4813]: I0317 09:31:21.370334 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85fc85fdfb-n469t" event={"ID":"0a52b24b-cc19-4720-a05b-961015362769","Type":"ContainerStarted","Data":"e4ff8b58e4f844a4867d183f3d7529361e570d7af0068b7705e4b87b56b07d24"} Mar 17 09:31:21 crc kubenswrapper[4813]: I0317 09:31:21.373344 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d927c167-348f-47f8-9860-b21dac3ca309","Type":"ContainerStarted","Data":"766a0ef148bc4a1cc73bfe82a1e3d829969f16ee65cb1fd39fd23aac1184be02"} Mar 17 09:31:21 crc kubenswrapper[4813]: I0317 09:31:21.381515 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7b8d767f8c-4ql8v" podStartSLOduration=3.30921625 podStartE2EDuration="28.381492323s" podCreationTimestamp="2026-03-17 09:30:53 +0000 UTC" firstStartedPulling="2026-03-17 09:30:54.045715991 +0000 UTC m=+1276.146519490" lastFinishedPulling="2026-03-17 09:31:19.117992064 +0000 UTC m=+1301.218795563" observedRunningTime="2026-03-17 09:31:21.378918602 +0000 UTC m=+1303.479722101" watchObservedRunningTime="2026-03-17 09:31:21.381492323 +0000 UTC m=+1303.482295822" Mar 17 09:31:21 crc kubenswrapper[4813]: I0317 09:31:21.388794 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5dc999f6f8-t94dl" event={"ID":"bf22966e-516d-40c2-975c-c3e41122b8d2","Type":"ContainerStarted","Data":"e760766c30a778b2fc07d2360ad4e72daf4e61dbf737479fef0099ccb1a9093e"} Mar 17 09:31:21 crc kubenswrapper[4813]: E0317 09:31:21.390421 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-27g9l" podUID="41809604-4a36-4453-b381-66b4c5cf1c43" Mar 17 09:31:21 crc kubenswrapper[4813]: I0317 09:31:21.481989 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-mfck7"] Mar 17 09:31:21 crc kubenswrapper[4813]: I0317 09:31:21.802519 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-b5495b5bd-g77kp"] Mar 17 09:31:21 crc kubenswrapper[4813]: W0317 09:31:21.849258 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podda12db2f_742c_4bf8_8c9e_3ee0868d0c63.slice/crio-bf28a2ce3451de4a157ee9315e929006bac898fa25cde725ee6f38f9ecec6311 WatchSource:0}: Error finding container bf28a2ce3451de4a157ee9315e929006bac898fa25cde725ee6f38f9ecec6311: Status 404 returned error can't find the container with id bf28a2ce3451de4a157ee9315e929006bac898fa25cde725ee6f38f9ecec6311 Mar 17 09:31:22 crc kubenswrapper[4813]: I0317 09:31:22.403994 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85fc85fdfb-n469t" event={"ID":"0a52b24b-cc19-4720-a05b-961015362769","Type":"ContainerStarted","Data":"77e6158b584698861562f3c366b15b1b09acecaf7c7eb2ee5edd3c9ddec18130"} Mar 17 09:31:22 crc kubenswrapper[4813]: I0317 09:31:22.404172 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85fc85fdfb-n469t" event={"ID":"0a52b24b-cc19-4720-a05b-961015362769","Type":"ContainerStarted","Data":"f4b4b9a07ddaca0201104a913247a2025ba56488f789d4d1b11072bb55859ee4"} Mar 17 09:31:22 crc kubenswrapper[4813]: I0317 09:31:22.410014 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b5495b5bd-g77kp" event={"ID":"da12db2f-742c-4bf8-8c9e-3ee0868d0c63","Type":"ContainerStarted","Data":"8c5f1c181c9c83af6c9eac398fcc89f4f7500de195959ec00e013725d482ce66"} Mar 17 09:31:22 crc kubenswrapper[4813]: I0317 09:31:22.410281 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b5495b5bd-g77kp" event={"ID":"da12db2f-742c-4bf8-8c9e-3ee0868d0c63","Type":"ContainerStarted","Data":"bf28a2ce3451de4a157ee9315e929006bac898fa25cde725ee6f38f9ecec6311"} Mar 17 09:31:22 crc kubenswrapper[4813]: I0317 09:31:22.412053 4813 generic.go:334] "Generic (PLEG): container finished" podID="de2cccf5-7797-4baf-bf2f-d6405959799f" containerID="b963f4bff251784bbbfa589d509487039ab20258d62b37f6cfe465f23047cca6" exitCode=0 Mar 17 09:31:22 crc kubenswrapper[4813]: I0317 09:31:22.412111 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" event={"ID":"de2cccf5-7797-4baf-bf2f-d6405959799f","Type":"ContainerDied","Data":"b963f4bff251784bbbfa589d509487039ab20258d62b37f6cfe465f23047cca6"} Mar 17 09:31:22 crc kubenswrapper[4813]: I0317 09:31:22.412127 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" event={"ID":"de2cccf5-7797-4baf-bf2f-d6405959799f","Type":"ContainerStarted","Data":"c37d2c93e6802dd7b7fbc5992f764c8658415869afaaf7df166092667ace066b"} Mar 17 09:31:22 crc kubenswrapper[4813]: I0317 09:31:22.418696 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5dc999f6f8-t94dl" event={"ID":"bf22966e-516d-40c2-975c-c3e41122b8d2","Type":"ContainerStarted","Data":"b39a68eed334735ed9f105585f3860123adc556ea6eb77fd5f15a9f8c7389419"} Mar 17 09:31:22 crc kubenswrapper[4813]: I0317 09:31:22.418738 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5dc999f6f8-t94dl" event={"ID":"bf22966e-516d-40c2-975c-c3e41122b8d2","Type":"ContainerStarted","Data":"04e5c4b0f2a78280497bce7330e313ab9414515665628dd6c0b436dbfbfd25e4"} Mar 17 09:31:22 crc kubenswrapper[4813]: I0317 09:31:22.424224 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"563f058b-5cf0-4f1f-91fb-c53c6088a9dc","Type":"ContainerStarted","Data":"a23facb1f8d187a1f432ea018fb2ca53da0f6eacfe0d1b993f572c13e0723bf0"} Mar 17 09:31:22 crc kubenswrapper[4813]: I0317 09:31:22.453354 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f01d7c15-58db-48db-84d5-e29b1edb2205","Type":"ContainerStarted","Data":"ee906bf72ea377aeeeefcb02192dc72de7b1b7e880ddafd917a4caaa9c6ffa19"} Mar 17 09:31:22 crc kubenswrapper[4813]: I0317 09:31:22.455226 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-85fc85fdfb-n469t" podStartSLOduration=23.455213912 podStartE2EDuration="23.455213912s" podCreationTimestamp="2026-03-17 09:30:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:31:22.43765391 +0000 UTC m=+1304.538457409" watchObservedRunningTime="2026-03-17 09:31:22.455213912 +0000 UTC m=+1304.556017411" Mar 17 09:31:22 crc kubenswrapper[4813]: I0317 09:31:22.462400 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5dc999f6f8-t94dl" podStartSLOduration=22.462387077 podStartE2EDuration="22.462387077s" podCreationTimestamp="2026-03-17 09:31:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:31:22.462010055 +0000 UTC m=+1304.562813554" watchObservedRunningTime="2026-03-17 09:31:22.462387077 +0000 UTC m=+1304.563190566" Mar 17 09:31:22 crc kubenswrapper[4813]: I0317 09:31:22.474311 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fdxt8" event={"ID":"00b72575-1230-444b-9c9b-6c2e2cc921f0","Type":"ContainerStarted","Data":"7d2252f8de6b1e8bafbbf5cf0d28de406e54ae318f14627bf88dbb57fc4ef071"} Mar 17 09:31:22 crc kubenswrapper[4813]: I0317 09:31:22.524312 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-fdxt8" podStartSLOduration=22.524291825 podStartE2EDuration="22.524291825s" podCreationTimestamp="2026-03-17 09:31:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:31:22.511563305 +0000 UTC m=+1304.612366804" watchObservedRunningTime="2026-03-17 09:31:22.524291825 +0000 UTC m=+1304.625095324" Mar 17 09:31:22 crc kubenswrapper[4813]: I0317 09:31:22.914037 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7ff5475cc9-crbwc" podUID="549ff45c-ad4e-4456-bf5a-eec6852665b9" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.133:5353: i/o timeout" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.039408 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-7d558687d7-q6tr2"] Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.041836 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.049000 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7d558687d7-q6tr2"] Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.053269 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.053687 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.193471 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-public-tls-certs\") pod \"neutron-7d558687d7-q6tr2\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.193535 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-ovndb-tls-certs\") pod \"neutron-7d558687d7-q6tr2\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.193560 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dtgff\" (UniqueName: \"kubernetes.io/projected/5aaee075-cdd4-4532-a651-539dfc8b6ac9-kube-api-access-dtgff\") pod \"neutron-7d558687d7-q6tr2\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.193578 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-combined-ca-bundle\") pod \"neutron-7d558687d7-q6tr2\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.193642 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-internal-tls-certs\") pod \"neutron-7d558687d7-q6tr2\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.193671 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-httpd-config\") pod \"neutron-7d558687d7-q6tr2\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.193695 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-config\") pod \"neutron-7d558687d7-q6tr2\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.295098 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-combined-ca-bundle\") pod \"neutron-7d558687d7-q6tr2\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.295183 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-internal-tls-certs\") pod \"neutron-7d558687d7-q6tr2\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.295213 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-httpd-config\") pod \"neutron-7d558687d7-q6tr2\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.295237 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-config\") pod \"neutron-7d558687d7-q6tr2\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.295298 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-public-tls-certs\") pod \"neutron-7d558687d7-q6tr2\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.295346 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-ovndb-tls-certs\") pod \"neutron-7d558687d7-q6tr2\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.295369 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dtgff\" (UniqueName: \"kubernetes.io/projected/5aaee075-cdd4-4532-a651-539dfc8b6ac9-kube-api-access-dtgff\") pod \"neutron-7d558687d7-q6tr2\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.310890 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-ovndb-tls-certs\") pod \"neutron-7d558687d7-q6tr2\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.313188 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-httpd-config\") pod \"neutron-7d558687d7-q6tr2\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.318115 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-config\") pod \"neutron-7d558687d7-q6tr2\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.320222 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dtgff\" (UniqueName: \"kubernetes.io/projected/5aaee075-cdd4-4532-a651-539dfc8b6ac9-kube-api-access-dtgff\") pod \"neutron-7d558687d7-q6tr2\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.321260 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-internal-tls-certs\") pod \"neutron-7d558687d7-q6tr2\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.325238 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-combined-ca-bundle\") pod \"neutron-7d558687d7-q6tr2\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.359483 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-public-tls-certs\") pod \"neutron-7d558687d7-q6tr2\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.367074 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.504939 4813 generic.go:334] "Generic (PLEG): container finished" podID="9c613a9b-36b2-437d-9c11-252d7d8ce910" containerID="3933cb26fc972a9b167b78a49896693ed6615d62ac3e9962d795dde203e9cd3f" exitCode=0 Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.505008 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hkmc6" event={"ID":"9c613a9b-36b2-437d-9c11-252d7d8ce910","Type":"ContainerDied","Data":"3933cb26fc972a9b167b78a49896693ed6615d62ac3e9962d795dde203e9cd3f"} Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.514259 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f01d7c15-58db-48db-84d5-e29b1edb2205","Type":"ContainerStarted","Data":"f545bbcc40b4a21f0118a88c62b42808441da4bceeb19202edc2e4932decd9d2"} Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.514410 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f01d7c15-58db-48db-84d5-e29b1edb2205" containerName="glance-log" containerID="cri-o://ee906bf72ea377aeeeefcb02192dc72de7b1b7e880ddafd917a4caaa9c6ffa19" gracePeriod=30 Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.514679 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f01d7c15-58db-48db-84d5-e29b1edb2205" containerName="glance-httpd" containerID="cri-o://f545bbcc40b4a21f0118a88c62b42808441da4bceeb19202edc2e4932decd9d2" gracePeriod=30 Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.518660 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7b8d767f8c-4ql8v" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.526063 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b5495b5bd-g77kp" event={"ID":"da12db2f-742c-4bf8-8c9e-3ee0868d0c63","Type":"ContainerStarted","Data":"c84fbe3bc0c39f7fff03f9ed586e1ddcb0e72ef85170e94045a5c1c7eabcb226"} Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.526916 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-b5495b5bd-g77kp" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.542786 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" event={"ID":"de2cccf5-7797-4baf-bf2f-d6405959799f","Type":"ContainerStarted","Data":"1b99d283d11cd2d81eba803093711bc717ff99fd015ca120f035046e76ba25dd"} Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.542890 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.550067 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="563f058b-5cf0-4f1f-91fb-c53c6088a9dc" containerName="glance-log" containerID="cri-o://a23facb1f8d187a1f432ea018fb2ca53da0f6eacfe0d1b993f572c13e0723bf0" gracePeriod=30 Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.550299 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"563f058b-5cf0-4f1f-91fb-c53c6088a9dc","Type":"ContainerStarted","Data":"c54d5cf6483e86ddc4f03258cfc9647987f653d9df2e0d792d725827ee504ea8"} Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.550716 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="563f058b-5cf0-4f1f-91fb-c53c6088a9dc" containerName="glance-httpd" containerID="cri-o://c54d5cf6483e86ddc4f03258cfc9647987f653d9df2e0d792d725827ee504ea8" gracePeriod=30 Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.576818 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=27.576772206 podStartE2EDuration="27.576772206s" podCreationTimestamp="2026-03-17 09:30:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:31:23.562214258 +0000 UTC m=+1305.663017757" watchObservedRunningTime="2026-03-17 09:31:23.576772206 +0000 UTC m=+1305.677575715" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.601336 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=28.601315507 podStartE2EDuration="28.601315507s" podCreationTimestamp="2026-03-17 09:30:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:31:23.5915198 +0000 UTC m=+1305.692323299" watchObservedRunningTime="2026-03-17 09:31:23.601315507 +0000 UTC m=+1305.702119006" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.628300 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" podStartSLOduration=3.628280926 podStartE2EDuration="3.628280926s" podCreationTimestamp="2026-03-17 09:31:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:31:23.621059048 +0000 UTC m=+1305.721862547" watchObservedRunningTime="2026-03-17 09:31:23.628280926 +0000 UTC m=+1305.729084425" Mar 17 09:31:23 crc kubenswrapper[4813]: I0317 09:31:23.668882 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-b5495b5bd-g77kp" podStartSLOduration=3.668865902 podStartE2EDuration="3.668865902s" podCreationTimestamp="2026-03-17 09:31:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:31:23.641523632 +0000 UTC m=+1305.742327131" watchObservedRunningTime="2026-03-17 09:31:23.668865902 +0000 UTC m=+1305.769669401" Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.564305 4813 generic.go:334] "Generic (PLEG): container finished" podID="563f058b-5cf0-4f1f-91fb-c53c6088a9dc" containerID="c54d5cf6483e86ddc4f03258cfc9647987f653d9df2e0d792d725827ee504ea8" exitCode=0 Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.564765 4813 generic.go:334] "Generic (PLEG): container finished" podID="563f058b-5cf0-4f1f-91fb-c53c6088a9dc" containerID="a23facb1f8d187a1f432ea018fb2ca53da0f6eacfe0d1b993f572c13e0723bf0" exitCode=143 Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.564807 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"563f058b-5cf0-4f1f-91fb-c53c6088a9dc","Type":"ContainerDied","Data":"c54d5cf6483e86ddc4f03258cfc9647987f653d9df2e0d792d725827ee504ea8"} Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.564830 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"563f058b-5cf0-4f1f-91fb-c53c6088a9dc","Type":"ContainerDied","Data":"a23facb1f8d187a1f432ea018fb2ca53da0f6eacfe0d1b993f572c13e0723bf0"} Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.564840 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"563f058b-5cf0-4f1f-91fb-c53c6088a9dc","Type":"ContainerDied","Data":"6e0e83bd78da580cc4b8d712b6069a9d023a31c09866a1bbbcabd82a6d042944"} Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.564848 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e0e83bd78da580cc4b8d712b6069a9d023a31c09866a1bbbcabd82a6d042944" Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.569134 4813 generic.go:334] "Generic (PLEG): container finished" podID="f01d7c15-58db-48db-84d5-e29b1edb2205" containerID="f545bbcc40b4a21f0118a88c62b42808441da4bceeb19202edc2e4932decd9d2" exitCode=0 Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.569156 4813 generic.go:334] "Generic (PLEG): container finished" podID="f01d7c15-58db-48db-84d5-e29b1edb2205" containerID="ee906bf72ea377aeeeefcb02192dc72de7b1b7e880ddafd917a4caaa9c6ffa19" exitCode=143 Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.569191 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f01d7c15-58db-48db-84d5-e29b1edb2205","Type":"ContainerDied","Data":"f545bbcc40b4a21f0118a88c62b42808441da4bceeb19202edc2e4932decd9d2"} Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.569212 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f01d7c15-58db-48db-84d5-e29b1edb2205","Type":"ContainerDied","Data":"ee906bf72ea377aeeeefcb02192dc72de7b1b7e880ddafd917a4caaa9c6ffa19"} Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.573678 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d927c167-348f-47f8-9860-b21dac3ca309","Type":"ContainerStarted","Data":"996ef8dc81ee850c6bcdf1656dc2e58131b9cf47f3b83174cf9f63eaddb3e604"} Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.583789 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.657113 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-7d558687d7-q6tr2"] Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.728273 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-httpd-run\") pod \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.728323 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-combined-ca-bundle\") pod \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.728353 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.728715 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-logs\") pod \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.728765 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-scripts\") pod \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.728838 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-config-data\") pod \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.728861 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8spbx\" (UniqueName: \"kubernetes.io/projected/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-kube-api-access-8spbx\") pod \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.728911 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-internal-tls-certs\") pod \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\" (UID: \"563f058b-5cf0-4f1f-91fb-c53c6088a9dc\") " Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.732984 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-logs" (OuterVolumeSpecName: "logs") pod "563f058b-5cf0-4f1f-91fb-c53c6088a9dc" (UID: "563f058b-5cf0-4f1f-91fb-c53c6088a9dc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.737556 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "563f058b-5cf0-4f1f-91fb-c53c6088a9dc" (UID: "563f058b-5cf0-4f1f-91fb-c53c6088a9dc"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.739116 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-scripts" (OuterVolumeSpecName: "scripts") pod "563f058b-5cf0-4f1f-91fb-c53c6088a9dc" (UID: "563f058b-5cf0-4f1f-91fb-c53c6088a9dc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.760569 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "563f058b-5cf0-4f1f-91fb-c53c6088a9dc" (UID: "563f058b-5cf0-4f1f-91fb-c53c6088a9dc"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.769769 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-kube-api-access-8spbx" (OuterVolumeSpecName: "kube-api-access-8spbx") pod "563f058b-5cf0-4f1f-91fb-c53c6088a9dc" (UID: "563f058b-5cf0-4f1f-91fb-c53c6088a9dc"). InnerVolumeSpecName "kube-api-access-8spbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.808697 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "563f058b-5cf0-4f1f-91fb-c53c6088a9dc" (UID: "563f058b-5cf0-4f1f-91fb-c53c6088a9dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.853729 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.853923 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.853946 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.853955 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-logs\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.853964 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.853972 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8spbx\" (UniqueName: \"kubernetes.io/projected/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-kube-api-access-8spbx\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.871408 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "563f058b-5cf0-4f1f-91fb-c53c6088a9dc" (UID: "563f058b-5cf0-4f1f-91fb-c53c6088a9dc"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.900310 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.904756 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-config-data" (OuterVolumeSpecName: "config-data") pod "563f058b-5cf0-4f1f-91fb-c53c6088a9dc" (UID: "563f058b-5cf0-4f1f-91fb-c53c6088a9dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.955151 4813 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.955176 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.955185 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/563f058b-5cf0-4f1f-91fb-c53c6088a9dc-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:24 crc kubenswrapper[4813]: I0317 09:31:24.987227 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hkmc6" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.165198 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2tnqs\" (UniqueName: \"kubernetes.io/projected/9c613a9b-36b2-437d-9c11-252d7d8ce910-kube-api-access-2tnqs\") pod \"9c613a9b-36b2-437d-9c11-252d7d8ce910\" (UID: \"9c613a9b-36b2-437d-9c11-252d7d8ce910\") " Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.165353 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c613a9b-36b2-437d-9c11-252d7d8ce910-scripts\") pod \"9c613a9b-36b2-437d-9c11-252d7d8ce910\" (UID: \"9c613a9b-36b2-437d-9c11-252d7d8ce910\") " Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.165455 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c613a9b-36b2-437d-9c11-252d7d8ce910-config-data\") pod \"9c613a9b-36b2-437d-9c11-252d7d8ce910\" (UID: \"9c613a9b-36b2-437d-9c11-252d7d8ce910\") " Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.165497 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c613a9b-36b2-437d-9c11-252d7d8ce910-logs\") pod \"9c613a9b-36b2-437d-9c11-252d7d8ce910\" (UID: \"9c613a9b-36b2-437d-9c11-252d7d8ce910\") " Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.165527 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c613a9b-36b2-437d-9c11-252d7d8ce910-combined-ca-bundle\") pod \"9c613a9b-36b2-437d-9c11-252d7d8ce910\" (UID: \"9c613a9b-36b2-437d-9c11-252d7d8ce910\") " Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.166804 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c613a9b-36b2-437d-9c11-252d7d8ce910-logs" (OuterVolumeSpecName: "logs") pod "9c613a9b-36b2-437d-9c11-252d7d8ce910" (UID: "9c613a9b-36b2-437d-9c11-252d7d8ce910"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.196849 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c613a9b-36b2-437d-9c11-252d7d8ce910-scripts" (OuterVolumeSpecName: "scripts") pod "9c613a9b-36b2-437d-9c11-252d7d8ce910" (UID: "9c613a9b-36b2-437d-9c11-252d7d8ce910"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.201881 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c613a9b-36b2-437d-9c11-252d7d8ce910-kube-api-access-2tnqs" (OuterVolumeSpecName: "kube-api-access-2tnqs") pod "9c613a9b-36b2-437d-9c11-252d7d8ce910" (UID: "9c613a9b-36b2-437d-9c11-252d7d8ce910"). InnerVolumeSpecName "kube-api-access-2tnqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.203679 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c613a9b-36b2-437d-9c11-252d7d8ce910-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c613a9b-36b2-437d-9c11-252d7d8ce910" (UID: "9c613a9b-36b2-437d-9c11-252d7d8ce910"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.216978 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c613a9b-36b2-437d-9c11-252d7d8ce910-config-data" (OuterVolumeSpecName: "config-data") pod "9c613a9b-36b2-437d-9c11-252d7d8ce910" (UID: "9c613a9b-36b2-437d-9c11-252d7d8ce910"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.294456 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9c613a9b-36b2-437d-9c11-252d7d8ce910-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.294492 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c613a9b-36b2-437d-9c11-252d7d8ce910-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.294503 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c613a9b-36b2-437d-9c11-252d7d8ce910-logs\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.294512 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c613a9b-36b2-437d-9c11-252d7d8ce910-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.294524 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2tnqs\" (UniqueName: \"kubernetes.io/projected/9c613a9b-36b2-437d-9c11-252d7d8ce910-kube-api-access-2tnqs\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.572806 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.598869 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-config-data\") pod \"f01d7c15-58db-48db-84d5-e29b1edb2205\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.599083 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f01d7c15-58db-48db-84d5-e29b1edb2205-httpd-run\") pod \"f01d7c15-58db-48db-84d5-e29b1edb2205\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.599121 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f01d7c15-58db-48db-84d5-e29b1edb2205-logs\") pod \"f01d7c15-58db-48db-84d5-e29b1edb2205\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.599332 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-scripts\") pod \"f01d7c15-58db-48db-84d5-e29b1edb2205\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.599426 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzhjg\" (UniqueName: \"kubernetes.io/projected/f01d7c15-58db-48db-84d5-e29b1edb2205-kube-api-access-fzhjg\") pod \"f01d7c15-58db-48db-84d5-e29b1edb2205\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.599532 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-public-tls-certs\") pod \"f01d7c15-58db-48db-84d5-e29b1edb2205\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.599719 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"f01d7c15-58db-48db-84d5-e29b1edb2205\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.599837 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-combined-ca-bundle\") pod \"f01d7c15-58db-48db-84d5-e29b1edb2205\" (UID: \"f01d7c15-58db-48db-84d5-e29b1edb2205\") " Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.600185 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f01d7c15-58db-48db-84d5-e29b1edb2205","Type":"ContainerDied","Data":"d925be3c7ce3ffb058eb969923deeb0f10c26ec78625dda621874a5dc90343e4"} Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.600425 4813 scope.go:117] "RemoveContainer" containerID="f545bbcc40b4a21f0118a88c62b42808441da4bceeb19202edc2e4932decd9d2" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.600238 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f01d7c15-58db-48db-84d5-e29b1edb2205-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f01d7c15-58db-48db-84d5-e29b1edb2205" (UID: "f01d7c15-58db-48db-84d5-e29b1edb2205"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.600858 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f01d7c15-58db-48db-84d5-e29b1edb2205-logs" (OuterVolumeSpecName: "logs") pod "f01d7c15-58db-48db-84d5-e29b1edb2205" (UID: "f01d7c15-58db-48db-84d5-e29b1edb2205"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.602192 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.609789 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d558687d7-q6tr2" event={"ID":"5aaee075-cdd4-4532-a651-539dfc8b6ac9","Type":"ContainerStarted","Data":"f4c6cd7b65fba4bb7d14a208ed65138752b3610ba036e75b33bc344a062202be"} Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.609829 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d558687d7-q6tr2" event={"ID":"5aaee075-cdd4-4532-a651-539dfc8b6ac9","Type":"ContainerStarted","Data":"8319e5f3fa73599000d2c08ec05ceacf5dcc9f68e9da36049d345fa482790c15"} Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.609839 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d558687d7-q6tr2" event={"ID":"5aaee075-cdd4-4532-a651-539dfc8b6ac9","Type":"ContainerStarted","Data":"cecfa89ed18ca4e06996dd45390bc4db8bdab3bd956c41a07ba7e58e30e9710d"} Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.612092 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.617223 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.620091 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f01d7c15-58db-48db-84d5-e29b1edb2205-kube-api-access-fzhjg" (OuterVolumeSpecName: "kube-api-access-fzhjg") pod "f01d7c15-58db-48db-84d5-e29b1edb2205" (UID: "f01d7c15-58db-48db-84d5-e29b1edb2205"). InnerVolumeSpecName "kube-api-access-fzhjg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.621331 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-hkmc6" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.622839 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-hkmc6" event={"ID":"9c613a9b-36b2-437d-9c11-252d7d8ce910","Type":"ContainerDied","Data":"ea9e78643d9a4c01a60114a74a0342ff8064039e5a3274d16ce37f85fec69a95"} Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.622881 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea9e78643d9a4c01a60114a74a0342ff8064039e5a3274d16ce37f85fec69a95" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.626053 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-scripts" (OuterVolumeSpecName: "scripts") pod "f01d7c15-58db-48db-84d5-e29b1edb2205" (UID: "f01d7c15-58db-48db-84d5-e29b1edb2205"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.626935 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "f01d7c15-58db-48db-84d5-e29b1edb2205" (UID: "f01d7c15-58db-48db-84d5-e29b1edb2205"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.684145 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-7d558687d7-q6tr2" podStartSLOduration=2.684129104 podStartE2EDuration="2.684129104s" podCreationTimestamp="2026-03-17 09:31:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:31:25.64202999 +0000 UTC m=+1307.742833500" watchObservedRunningTime="2026-03-17 09:31:25.684129104 +0000 UTC m=+1307.784932603" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.691179 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-6cc74cf554-dmnmf"] Mar 17 09:31:25 crc kubenswrapper[4813]: E0317 09:31:25.691528 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="563f058b-5cf0-4f1f-91fb-c53c6088a9dc" containerName="glance-httpd" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.691539 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="563f058b-5cf0-4f1f-91fb-c53c6088a9dc" containerName="glance-httpd" Mar 17 09:31:25 crc kubenswrapper[4813]: E0317 09:31:25.691550 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f01d7c15-58db-48db-84d5-e29b1edb2205" containerName="glance-httpd" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.691555 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f01d7c15-58db-48db-84d5-e29b1edb2205" containerName="glance-httpd" Mar 17 09:31:25 crc kubenswrapper[4813]: E0317 09:31:25.691564 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f01d7c15-58db-48db-84d5-e29b1edb2205" containerName="glance-log" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.691570 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f01d7c15-58db-48db-84d5-e29b1edb2205" containerName="glance-log" Mar 17 09:31:25 crc kubenswrapper[4813]: E0317 09:31:25.691579 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="563f058b-5cf0-4f1f-91fb-c53c6088a9dc" containerName="glance-log" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.691585 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="563f058b-5cf0-4f1f-91fb-c53c6088a9dc" containerName="glance-log" Mar 17 09:31:25 crc kubenswrapper[4813]: E0317 09:31:25.691613 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c613a9b-36b2-437d-9c11-252d7d8ce910" containerName="placement-db-sync" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.692590 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c613a9b-36b2-437d-9c11-252d7d8ce910" containerName="placement-db-sync" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.692829 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c613a9b-36b2-437d-9c11-252d7d8ce910" containerName="placement-db-sync" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.692857 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f01d7c15-58db-48db-84d5-e29b1edb2205" containerName="glance-log" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.692868 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="563f058b-5cf0-4f1f-91fb-c53c6088a9dc" containerName="glance-httpd" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.692887 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="563f058b-5cf0-4f1f-91fb-c53c6088a9dc" containerName="glance-log" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.692895 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f01d7c15-58db-48db-84d5-e29b1edb2205" containerName="glance-httpd" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.693756 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.702157 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb24e4e7-5911-4a4b-bff0-e0de7087be65-logs\") pod \"placement-6cc74cf554-dmnmf\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.702193 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-config-data\") pod \"placement-6cc74cf554-dmnmf\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.702220 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-internal-tls-certs\") pod \"placement-6cc74cf554-dmnmf\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.702260 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-combined-ca-bundle\") pod \"placement-6cc74cf554-dmnmf\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.702278 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-scripts\") pod \"placement-6cc74cf554-dmnmf\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.702328 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhbb9\" (UniqueName: \"kubernetes.io/projected/bb24e4e7-5911-4a4b-bff0-e0de7087be65-kube-api-access-jhbb9\") pod \"placement-6cc74cf554-dmnmf\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.702397 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-public-tls-certs\") pod \"placement-6cc74cf554-dmnmf\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.702517 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.702530 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f01d7c15-58db-48db-84d5-e29b1edb2205-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.702539 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f01d7c15-58db-48db-84d5-e29b1edb2205-logs\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.702547 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.702555 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzhjg\" (UniqueName: \"kubernetes.io/projected/f01d7c15-58db-48db-84d5-e29b1edb2205-kube-api-access-fzhjg\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.706117 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-5l9nj" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.706393 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.706659 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.706772 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.706876 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.717796 4813 scope.go:117] "RemoveContainer" containerID="ee906bf72ea377aeeeefcb02192dc72de7b1b7e880ddafd917a4caaa9c6ffa19" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.739143 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6cc74cf554-dmnmf"] Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.739839 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-config-data" (OuterVolumeSpecName: "config-data") pod "f01d7c15-58db-48db-84d5-e29b1edb2205" (UID: "f01d7c15-58db-48db-84d5-e29b1edb2205"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.747107 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.763797 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f01d7c15-58db-48db-84d5-e29b1edb2205" (UID: "f01d7c15-58db-48db-84d5-e29b1edb2205"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.776025 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.797658 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.798099 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f01d7c15-58db-48db-84d5-e29b1edb2205" (UID: "f01d7c15-58db-48db-84d5-e29b1edb2205"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.803941 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb24e4e7-5911-4a4b-bff0-e0de7087be65-logs\") pod \"placement-6cc74cf554-dmnmf\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.803980 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-config-data\") pod \"placement-6cc74cf554-dmnmf\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.804004 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-internal-tls-certs\") pod \"placement-6cc74cf554-dmnmf\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.804038 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-combined-ca-bundle\") pod \"placement-6cc74cf554-dmnmf\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.804057 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-scripts\") pod \"placement-6cc74cf554-dmnmf\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.804095 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhbb9\" (UniqueName: \"kubernetes.io/projected/bb24e4e7-5911-4a4b-bff0-e0de7087be65-kube-api-access-jhbb9\") pod \"placement-6cc74cf554-dmnmf\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.804144 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-public-tls-certs\") pod \"placement-6cc74cf554-dmnmf\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.804224 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.804404 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.804784 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb24e4e7-5911-4a4b-bff0-e0de7087be65-logs\") pod \"placement-6cc74cf554-dmnmf\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.810202 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.810405 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-public-tls-certs\") pod \"placement-6cc74cf554-dmnmf\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.810498 4813 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f01d7c15-58db-48db-84d5-e29b1edb2205-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.812059 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-combined-ca-bundle\") pod \"placement-6cc74cf554-dmnmf\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.819931 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-scripts\") pod \"placement-6cc74cf554-dmnmf\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.820101 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-internal-tls-certs\") pod \"placement-6cc74cf554-dmnmf\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.820497 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-config-data\") pod \"placement-6cc74cf554-dmnmf\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.829187 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.830746 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.839134 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.839375 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.840489 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhbb9\" (UniqueName: \"kubernetes.io/projected/bb24e4e7-5911-4a4b-bff0-e0de7087be65-kube-api-access-jhbb9\") pod \"placement-6cc74cf554-dmnmf\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.855674 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.917553 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.917790 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.918950 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc74e31f-b87b-4888-b99f-342c2f72f6b7-logs\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.919030 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.919122 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fc74e31f-b87b-4888-b99f-342c2f72f6b7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.919161 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sh2zr\" (UniqueName: \"kubernetes.io/projected/fc74e31f-b87b-4888-b99f-342c2f72f6b7-kube-api-access-sh2zr\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.919217 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.922471 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.967687 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.981752 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.988358 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.989680 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.993914 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 17 09:31:25 crc kubenswrapper[4813]: I0317 09:31:25.994058 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.008374 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.027818 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-scripts\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.027873 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-config-data\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.027913 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc74e31f-b87b-4888-b99f-342c2f72f6b7-logs\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.027934 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.027959 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sh2zr\" (UniqueName: \"kubernetes.io/projected/fc74e31f-b87b-4888-b99f-342c2f72f6b7-kube-api-access-sh2zr\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.027973 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fc74e31f-b87b-4888-b99f-342c2f72f6b7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.028007 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.028030 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.028050 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.028067 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.028092 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.028137 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52qfr\" (UniqueName: \"kubernetes.io/projected/a1c5331f-3325-4bab-a0de-016366febfcd-kube-api-access-52qfr\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.028153 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1c5331f-3325-4bab-a0de-016366febfcd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.028169 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.028185 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.028198 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c5331f-3325-4bab-a0de-016366febfcd-logs\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.028636 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc74e31f-b87b-4888-b99f-342c2f72f6b7-logs\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.033036 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.033608 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fc74e31f-b87b-4888-b99f-342c2f72f6b7-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.033886 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.039551 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.040572 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-config-data\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.049980 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.065108 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-scripts\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.066330 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sh2zr\" (UniqueName: \"kubernetes.io/projected/fc74e31f-b87b-4888-b99f-342c2f72f6b7-kube-api-access-sh2zr\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.072461 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.131449 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-scripts\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.131512 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-config-data\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.131586 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.131625 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.131672 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52qfr\" (UniqueName: \"kubernetes.io/projected/a1c5331f-3325-4bab-a0de-016366febfcd-kube-api-access-52qfr\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.131690 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1c5331f-3325-4bab-a0de-016366febfcd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.131708 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.131728 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c5331f-3325-4bab-a0de-016366febfcd-logs\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.132111 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c5331f-3325-4bab-a0de-016366febfcd-logs\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.133615 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.137066 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1c5331f-3325-4bab-a0de-016366febfcd-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.140126 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-scripts\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.141990 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.150245 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52qfr\" (UniqueName: \"kubernetes.io/projected/a1c5331f-3325-4bab-a0de-016366febfcd-kube-api-access-52qfr\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.153762 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-config-data\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.154090 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.158186 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.228260 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.331155 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.654977 4813 generic.go:334] "Generic (PLEG): container finished" podID="00b72575-1230-444b-9c9b-6c2e2cc921f0" containerID="7d2252f8de6b1e8bafbbf5cf0d28de406e54ae318f14627bf88dbb57fc4ef071" exitCode=0 Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.655610 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fdxt8" event={"ID":"00b72575-1230-444b-9c9b-6c2e2cc921f0","Type":"ContainerDied","Data":"7d2252f8de6b1e8bafbbf5cf0d28de406e54ae318f14627bf88dbb57fc4ef071"} Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.695831 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-6cc74cf554-dmnmf"] Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.780134 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="563f058b-5cf0-4f1f-91fb-c53c6088a9dc" path="/var/lib/kubelet/pods/563f058b-5cf0-4f1f-91fb-c53c6088a9dc/volumes" Mar 17 09:31:26 crc kubenswrapper[4813]: I0317 09:31:26.781155 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f01d7c15-58db-48db-84d5-e29b1edb2205" path="/var/lib/kubelet/pods/f01d7c15-58db-48db-84d5-e29b1edb2205/volumes" Mar 17 09:31:27 crc kubenswrapper[4813]: I0317 09:31:27.022526 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 09:31:27 crc kubenswrapper[4813]: I0317 09:31:27.256007 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 09:31:27 crc kubenswrapper[4813]: I0317 09:31:27.711543 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fc74e31f-b87b-4888-b99f-342c2f72f6b7","Type":"ContainerStarted","Data":"52d6e9df3f7cb678ef46427439a32bd53667534a016fabfd27a7f8359286cdf3"} Mar 17 09:31:27 crc kubenswrapper[4813]: I0317 09:31:27.713067 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a1c5331f-3325-4bab-a0de-016366febfcd","Type":"ContainerStarted","Data":"6b71041d6c374849f3739c6a41ddf51e58df25ee101febcff6c3077ccad61165"} Mar 17 09:31:27 crc kubenswrapper[4813]: I0317 09:31:27.720512 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6cc74cf554-dmnmf" event={"ID":"bb24e4e7-5911-4a4b-bff0-e0de7087be65","Type":"ContainerStarted","Data":"f54d815fe296100a77eea0b4c0830a67b317b2b603abccbc9fa035e4ca28c4bb"} Mar 17 09:31:27 crc kubenswrapper[4813]: I0317 09:31:27.720550 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6cc74cf554-dmnmf" event={"ID":"bb24e4e7-5911-4a4b-bff0-e0de7087be65","Type":"ContainerStarted","Data":"0df61a5ea2fb36b21335d446d9790e9116aa82a0d3dfb4105c46092354b3a6d2"} Mar 17 09:31:27 crc kubenswrapper[4813]: I0317 09:31:27.720563 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6cc74cf554-dmnmf" event={"ID":"bb24e4e7-5911-4a4b-bff0-e0de7087be65","Type":"ContainerStarted","Data":"801f34584e1cf811b7cda08ea8a4678f0f527fd3d3b37b10c0ae9e0211c56381"} Mar 17 09:31:27 crc kubenswrapper[4813]: I0317 09:31:27.752715 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-6cc74cf554-dmnmf" podStartSLOduration=2.752696533 podStartE2EDuration="2.752696533s" podCreationTimestamp="2026-03-17 09:31:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:31:27.745129744 +0000 UTC m=+1309.845933243" watchObservedRunningTime="2026-03-17 09:31:27.752696533 +0000 UTC m=+1309.853500032" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.273549 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.302826 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-credential-keys\") pod \"00b72575-1230-444b-9c9b-6c2e2cc921f0\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.302902 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-config-data\") pod \"00b72575-1230-444b-9c9b-6c2e2cc921f0\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.303294 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-scripts\") pod \"00b72575-1230-444b-9c9b-6c2e2cc921f0\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.303318 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-fernet-keys\") pod \"00b72575-1230-444b-9c9b-6c2e2cc921f0\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.303342 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tl46r\" (UniqueName: \"kubernetes.io/projected/00b72575-1230-444b-9c9b-6c2e2cc921f0-kube-api-access-tl46r\") pod \"00b72575-1230-444b-9c9b-6c2e2cc921f0\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.303443 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-combined-ca-bundle\") pod \"00b72575-1230-444b-9c9b-6c2e2cc921f0\" (UID: \"00b72575-1230-444b-9c9b-6c2e2cc921f0\") " Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.316561 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "00b72575-1230-444b-9c9b-6c2e2cc921f0" (UID: "00b72575-1230-444b-9c9b-6c2e2cc921f0"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.317793 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00b72575-1230-444b-9c9b-6c2e2cc921f0-kube-api-access-tl46r" (OuterVolumeSpecName: "kube-api-access-tl46r") pod "00b72575-1230-444b-9c9b-6c2e2cc921f0" (UID: "00b72575-1230-444b-9c9b-6c2e2cc921f0"). InnerVolumeSpecName "kube-api-access-tl46r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.318372 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tl46r\" (UniqueName: \"kubernetes.io/projected/00b72575-1230-444b-9c9b-6c2e2cc921f0-kube-api-access-tl46r\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.318400 4813 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-credential-keys\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.340563 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-config-data" (OuterVolumeSpecName: "config-data") pod "00b72575-1230-444b-9c9b-6c2e2cc921f0" (UID: "00b72575-1230-444b-9c9b-6c2e2cc921f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.352334 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "00b72575-1230-444b-9c9b-6c2e2cc921f0" (UID: "00b72575-1230-444b-9c9b-6c2e2cc921f0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.354832 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-scripts" (OuterVolumeSpecName: "scripts") pod "00b72575-1230-444b-9c9b-6c2e2cc921f0" (UID: "00b72575-1230-444b-9c9b-6c2e2cc921f0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.367775 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "00b72575-1230-444b-9c9b-6c2e2cc921f0" (UID: "00b72575-1230-444b-9c9b-6c2e2cc921f0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.420399 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.420428 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.420437 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.420445 4813 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/00b72575-1230-444b-9c9b-6c2e2cc921f0-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.736516 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-fdxt8" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.745459 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a1c5331f-3325-4bab-a0de-016366febfcd","Type":"ContainerStarted","Data":"b571f7ac595fd0a9269bdb93504580575de28b2f8b55d4c8d29eb60d02fad58e"} Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.745496 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-fdxt8" event={"ID":"00b72575-1230-444b-9c9b-6c2e2cc921f0","Type":"ContainerDied","Data":"95d4fd8e2925edcb3f6c488c377d44212e999965266b1e21641072d605c009c9"} Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.745508 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="95d4fd8e2925edcb3f6c488c377d44212e999965266b1e21641072d605c009c9" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.745521 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.745533 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fc74e31f-b87b-4888-b99f-342c2f72f6b7","Type":"ContainerStarted","Data":"20afc3b693977dd502ec4c0428f5b9fd7987ff5309f74555d32af8f5c64e9012"} Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.745544 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.822425 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-7fb6796ddc-58j2l"] Mar 17 09:31:28 crc kubenswrapper[4813]: E0317 09:31:28.830232 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00b72575-1230-444b-9c9b-6c2e2cc921f0" containerName="keystone-bootstrap" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.830265 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="00b72575-1230-444b-9c9b-6c2e2cc921f0" containerName="keystone-bootstrap" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.830466 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="00b72575-1230-444b-9c9b-6c2e2cc921f0" containerName="keystone-bootstrap" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.831033 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.836192 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.836388 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.836200 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-tnvpg" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.836395 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.836762 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.836939 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.839005 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7fb6796ddc-58j2l"] Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.929580 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd0582a1-002f-452c-828a-406a4d945f28-internal-tls-certs\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.929697 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd0582a1-002f-452c-828a-406a4d945f28-scripts\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.929751 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd0582a1-002f-452c-828a-406a4d945f28-public-tls-certs\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.929817 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd0582a1-002f-452c-828a-406a4d945f28-config-data\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.929862 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd0582a1-002f-452c-828a-406a4d945f28-combined-ca-bundle\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.929905 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdw8b\" (UniqueName: \"kubernetes.io/projected/fd0582a1-002f-452c-828a-406a4d945f28-kube-api-access-sdw8b\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.930251 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fd0582a1-002f-452c-828a-406a4d945f28-credential-keys\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:28 crc kubenswrapper[4813]: I0317 09:31:28.930291 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fd0582a1-002f-452c-828a-406a4d945f28-fernet-keys\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.032044 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd0582a1-002f-452c-828a-406a4d945f28-config-data\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.032117 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd0582a1-002f-452c-828a-406a4d945f28-combined-ca-bundle\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.032167 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdw8b\" (UniqueName: \"kubernetes.io/projected/fd0582a1-002f-452c-828a-406a4d945f28-kube-api-access-sdw8b\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.032196 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fd0582a1-002f-452c-828a-406a4d945f28-credential-keys\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.032227 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fd0582a1-002f-452c-828a-406a4d945f28-fernet-keys\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.032262 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd0582a1-002f-452c-828a-406a4d945f28-internal-tls-certs\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.032294 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd0582a1-002f-452c-828a-406a4d945f28-scripts\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.032343 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd0582a1-002f-452c-828a-406a4d945f28-public-tls-certs\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.040405 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd0582a1-002f-452c-828a-406a4d945f28-public-tls-certs\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.041069 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd0582a1-002f-452c-828a-406a4d945f28-combined-ca-bundle\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.041450 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/fd0582a1-002f-452c-828a-406a4d945f28-credential-keys\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.041825 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fd0582a1-002f-452c-828a-406a4d945f28-scripts\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.044500 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fd0582a1-002f-452c-828a-406a4d945f28-internal-tls-certs\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.044763 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/fd0582a1-002f-452c-828a-406a4d945f28-fernet-keys\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.047265 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd0582a1-002f-452c-828a-406a4d945f28-config-data\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.072984 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdw8b\" (UniqueName: \"kubernetes.io/projected/fd0582a1-002f-452c-828a-406a4d945f28-kube-api-access-sdw8b\") pod \"keystone-7fb6796ddc-58j2l\" (UID: \"fd0582a1-002f-452c-828a-406a4d945f28\") " pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.206385 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.745028 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-7fb6796ddc-58j2l"] Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.764575 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a1c5331f-3325-4bab-a0de-016366febfcd","Type":"ContainerStarted","Data":"6a8330860efcd7dff1fc03e67f918941c9765c0844a3b969348cd2ca24a98307"} Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.778733 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fc74e31f-b87b-4888-b99f-342c2f72f6b7","Type":"ContainerStarted","Data":"12e6c7f140dd842351032cd69a0fd18874b77ca813d3ea3f119f9553b1b4b862"} Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.796188 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.796170361 podStartE2EDuration="4.796170361s" podCreationTimestamp="2026-03-17 09:31:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:31:29.788105458 +0000 UTC m=+1311.888908957" watchObservedRunningTime="2026-03-17 09:31:29.796170361 +0000 UTC m=+1311.896973860" Mar 17 09:31:29 crc kubenswrapper[4813]: I0317 09:31:29.816068 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.816051347 podStartE2EDuration="4.816051347s" podCreationTimestamp="2026-03-17 09:31:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:31:29.812945859 +0000 UTC m=+1311.913749358" watchObservedRunningTime="2026-03-17 09:31:29.816051347 +0000 UTC m=+1311.916854836" Mar 17 09:31:30 crc kubenswrapper[4813]: I0317 09:31:30.305802 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:31:30 crc kubenswrapper[4813]: I0317 09:31:30.306127 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:31:30 crc kubenswrapper[4813]: I0317 09:31:30.491356 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:30 crc kubenswrapper[4813]: I0317 09:31:30.492268 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:30 crc kubenswrapper[4813]: I0317 09:31:30.795397 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7fb6796ddc-58j2l" event={"ID":"fd0582a1-002f-452c-828a-406a4d945f28","Type":"ContainerStarted","Data":"317a1d8420b15787fb7f5fb9b61805a15c8bb3fa0cc0a5430e67b20e47f390be"} Mar 17 09:31:30 crc kubenswrapper[4813]: I0317 09:31:30.795433 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-7fb6796ddc-58j2l" event={"ID":"fd0582a1-002f-452c-828a-406a4d945f28","Type":"ContainerStarted","Data":"714bc3446c74fbf8a161da2bf1deda3c0dcbf4cbebb81c79d6ade8edefdc77ec"} Mar 17 09:31:30 crc kubenswrapper[4813]: I0317 09:31:30.796613 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:31:30 crc kubenswrapper[4813]: I0317 09:31:30.830458 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-7fb6796ddc-58j2l" podStartSLOduration=2.830275714 podStartE2EDuration="2.830275714s" podCreationTimestamp="2026-03-17 09:31:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:31:30.818085291 +0000 UTC m=+1312.918888790" watchObservedRunningTime="2026-03-17 09:31:30.830275714 +0000 UTC m=+1312.931079213" Mar 17 09:31:30 crc kubenswrapper[4813]: I0317 09:31:30.893964 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:30 crc kubenswrapper[4813]: I0317 09:31:30.950181 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-6wc77"] Mar 17 09:31:30 crc kubenswrapper[4813]: I0317 09:31:30.950442 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" podUID="1596452f-363d-4d92-82de-f96f1f83993a" containerName="dnsmasq-dns" containerID="cri-o://77b92a89118e3453f1d95103179cd5750125b99f7124117d4357e999f4f5584b" gracePeriod=10 Mar 17 09:31:31 crc kubenswrapper[4813]: I0317 09:31:31.805044 4813 generic.go:334] "Generic (PLEG): container finished" podID="1596452f-363d-4d92-82de-f96f1f83993a" containerID="77b92a89118e3453f1d95103179cd5750125b99f7124117d4357e999f4f5584b" exitCode=0 Mar 17 09:31:31 crc kubenswrapper[4813]: I0317 09:31:31.805135 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" event={"ID":"1596452f-363d-4d92-82de-f96f1f83993a","Type":"ContainerDied","Data":"77b92a89118e3453f1d95103179cd5750125b99f7124117d4357e999f4f5584b"} Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.685777 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.755780 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-ovsdbserver-sb\") pod \"1596452f-363d-4d92-82de-f96f1f83993a\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.755854 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-config\") pod \"1596452f-363d-4d92-82de-f96f1f83993a\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.755900 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-dns-swift-storage-0\") pod \"1596452f-363d-4d92-82de-f96f1f83993a\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.755950 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-dns-svc\") pod \"1596452f-363d-4d92-82de-f96f1f83993a\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.755985 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-ovsdbserver-nb\") pod \"1596452f-363d-4d92-82de-f96f1f83993a\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.756070 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g5vfq\" (UniqueName: \"kubernetes.io/projected/1596452f-363d-4d92-82de-f96f1f83993a-kube-api-access-g5vfq\") pod \"1596452f-363d-4d92-82de-f96f1f83993a\" (UID: \"1596452f-363d-4d92-82de-f96f1f83993a\") " Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.770282 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1596452f-363d-4d92-82de-f96f1f83993a-kube-api-access-g5vfq" (OuterVolumeSpecName: "kube-api-access-g5vfq") pod "1596452f-363d-4d92-82de-f96f1f83993a" (UID: "1596452f-363d-4d92-82de-f96f1f83993a"). InnerVolumeSpecName "kube-api-access-g5vfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.827296 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "1596452f-363d-4d92-82de-f96f1f83993a" (UID: "1596452f-363d-4d92-82de-f96f1f83993a"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.828030 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "1596452f-363d-4d92-82de-f96f1f83993a" (UID: "1596452f-363d-4d92-82de-f96f1f83993a"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.828080 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1596452f-363d-4d92-82de-f96f1f83993a" (UID: "1596452f-363d-4d92-82de-f96f1f83993a"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.835076 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1596452f-363d-4d92-82de-f96f1f83993a" (UID: "1596452f-363d-4d92-82de-f96f1f83993a"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.838629 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-config" (OuterVolumeSpecName: "config") pod "1596452f-363d-4d92-82de-f96f1f83993a" (UID: "1596452f-363d-4d92-82de-f96f1f83993a"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.839714 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d927c167-348f-47f8-9860-b21dac3ca309","Type":"ContainerStarted","Data":"faad4b4a7c4535789ba6e076f9588c4385a53196f95b284dcf17a03fd3bc815f"} Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.842331 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" event={"ID":"1596452f-363d-4d92-82de-f96f1f83993a","Type":"ContainerDied","Data":"b393b89dea2a6d9b43f6704b679c1708b47ccc5c4c66126d4a6d9850df1b5086"} Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.842363 4813 scope.go:117] "RemoveContainer" containerID="77b92a89118e3453f1d95103179cd5750125b99f7124117d4357e999f4f5584b" Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.842479 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8b5c85b87-6wc77" Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.857917 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.857945 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.857954 4813 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.857963 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.857972 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/1596452f-363d-4d92-82de-f96f1f83993a-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.857982 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g5vfq\" (UniqueName: \"kubernetes.io/projected/1596452f-363d-4d92-82de-f96f1f83993a-kube-api-access-g5vfq\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.875530 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-6wc77"] Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.882738 4813 scope.go:117] "RemoveContainer" containerID="4fe25e5c2f86364582e41039c2cd8fbf6c091504a790e69bcbb74f5da0ff3642" Mar 17 09:31:34 crc kubenswrapper[4813]: I0317 09:31:34.882745 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8b5c85b87-6wc77"] Mar 17 09:31:35 crc kubenswrapper[4813]: I0317 09:31:35.854424 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-27g9l" event={"ID":"41809604-4a36-4453-b381-66b4c5cf1c43","Type":"ContainerStarted","Data":"c0bcfc07898a4177c7b334fd23c56e6fbd4fa2af6b58ab9ba3a9942136ae713b"} Mar 17 09:31:35 crc kubenswrapper[4813]: I0317 09:31:35.859722 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-mxghx" event={"ID":"5c6cd309-922e-4be6-b2d9-12d2aca54b78","Type":"ContainerStarted","Data":"2e13e45a5cbf8e2e08800c4276301a0d39702765d34df2bcb61d04f52b11d8a9"} Mar 17 09:31:35 crc kubenswrapper[4813]: I0317 09:31:35.896294 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-27g9l" podStartSLOduration=4.173044774 podStartE2EDuration="46.896272934s" podCreationTimestamp="2026-03-17 09:30:49 +0000 UTC" firstStartedPulling="2026-03-17 09:30:51.774819369 +0000 UTC m=+1273.875622868" lastFinishedPulling="2026-03-17 09:31:34.498047529 +0000 UTC m=+1316.598851028" observedRunningTime="2026-03-17 09:31:35.888978914 +0000 UTC m=+1317.989782423" watchObservedRunningTime="2026-03-17 09:31:35.896272934 +0000 UTC m=+1317.997076453" Mar 17 09:31:35 crc kubenswrapper[4813]: I0317 09:31:35.904520 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-mxghx" podStartSLOduration=2.191819524 podStartE2EDuration="45.904498623s" podCreationTimestamp="2026-03-17 09:30:50 +0000 UTC" firstStartedPulling="2026-03-17 09:30:51.713041096 +0000 UTC m=+1273.813844595" lastFinishedPulling="2026-03-17 09:31:35.425720155 +0000 UTC m=+1317.526523694" observedRunningTime="2026-03-17 09:31:35.901425286 +0000 UTC m=+1318.002228785" watchObservedRunningTime="2026-03-17 09:31:35.904498623 +0000 UTC m=+1318.005302122" Mar 17 09:31:36 crc kubenswrapper[4813]: I0317 09:31:36.159656 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 17 09:31:36 crc kubenswrapper[4813]: I0317 09:31:36.164035 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 17 09:31:36 crc kubenswrapper[4813]: I0317 09:31:36.202523 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 17 09:31:36 crc kubenswrapper[4813]: I0317 09:31:36.216023 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 17 09:31:36 crc kubenswrapper[4813]: I0317 09:31:36.331723 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 17 09:31:36 crc kubenswrapper[4813]: I0317 09:31:36.331773 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 17 09:31:36 crc kubenswrapper[4813]: I0317 09:31:36.360367 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 17 09:31:36 crc kubenswrapper[4813]: I0317 09:31:36.370796 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 17 09:31:36 crc kubenswrapper[4813]: I0317 09:31:36.746353 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1596452f-363d-4d92-82de-f96f1f83993a" path="/var/lib/kubelet/pods/1596452f-363d-4d92-82de-f96f1f83993a/volumes" Mar 17 09:31:36 crc kubenswrapper[4813]: I0317 09:31:36.879490 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 17 09:31:36 crc kubenswrapper[4813]: I0317 09:31:36.879554 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 17 09:31:36 crc kubenswrapper[4813]: I0317 09:31:36.879573 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 17 09:31:36 crc kubenswrapper[4813]: I0317 09:31:36.879790 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 17 09:31:38 crc kubenswrapper[4813]: I0317 09:31:38.902689 4813 generic.go:334] "Generic (PLEG): container finished" podID="5c6cd309-922e-4be6-b2d9-12d2aca54b78" containerID="2e13e45a5cbf8e2e08800c4276301a0d39702765d34df2bcb61d04f52b11d8a9" exitCode=0 Mar 17 09:31:38 crc kubenswrapper[4813]: I0317 09:31:38.903086 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-mxghx" event={"ID":"5c6cd309-922e-4be6-b2d9-12d2aca54b78","Type":"ContainerDied","Data":"2e13e45a5cbf8e2e08800c4276301a0d39702765d34df2bcb61d04f52b11d8a9"} Mar 17 09:31:39 crc kubenswrapper[4813]: I0317 09:31:39.122700 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 17 09:31:39 crc kubenswrapper[4813]: I0317 09:31:39.122795 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 17 09:31:39 crc kubenswrapper[4813]: I0317 09:31:39.125937 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 17 09:31:39 crc kubenswrapper[4813]: I0317 09:31:39.252780 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 17 09:31:39 crc kubenswrapper[4813]: I0317 09:31:39.252904 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 17 09:31:39 crc kubenswrapper[4813]: I0317 09:31:39.256992 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 17 09:31:40 crc kubenswrapper[4813]: I0317 09:31:40.307678 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-85fc85fdfb-n469t" podUID="0a52b24b-cc19-4720-a05b-961015362769" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.156:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.156:8443: connect: connection refused" Mar 17 09:31:40 crc kubenswrapper[4813]: I0317 09:31:40.492536 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5dc999f6f8-t94dl" podUID="bf22966e-516d-40c2-975c-c3e41122b8d2" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.157:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.157:8443: connect: connection refused" Mar 17 09:31:40 crc kubenswrapper[4813]: I0317 09:31:40.945083 4813 generic.go:334] "Generic (PLEG): container finished" podID="41809604-4a36-4453-b381-66b4c5cf1c43" containerID="c0bcfc07898a4177c7b334fd23c56e6fbd4fa2af6b58ab9ba3a9942136ae713b" exitCode=0 Mar 17 09:31:40 crc kubenswrapper[4813]: I0317 09:31:40.945129 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-27g9l" event={"ID":"41809604-4a36-4453-b381-66b4c5cf1c43","Type":"ContainerDied","Data":"c0bcfc07898a4177c7b334fd23c56e6fbd4fa2af6b58ab9ba3a9942136ae713b"} Mar 17 09:31:43 crc kubenswrapper[4813]: I0317 09:31:43.980879 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-mxghx" event={"ID":"5c6cd309-922e-4be6-b2d9-12d2aca54b78","Type":"ContainerDied","Data":"d7ee9fb0dab48c1954c52b2160ee734071056c2ba77403f85c1e40679cdd5170"} Mar 17 09:31:43 crc kubenswrapper[4813]: I0317 09:31:43.981358 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7ee9fb0dab48c1954c52b2160ee734071056c2ba77403f85c1e40679cdd5170" Mar 17 09:31:43 crc kubenswrapper[4813]: I0317 09:31:43.983949 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-27g9l" event={"ID":"41809604-4a36-4453-b381-66b4c5cf1c43","Type":"ContainerDied","Data":"4bd5f71c52301a2b0cf8e50675e7fb31caf20086f6fdd0ee80fbd5f75e50fe39"} Mar 17 09:31:43 crc kubenswrapper[4813]: I0317 09:31:43.983983 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bd5f71c52301a2b0cf8e50675e7fb31caf20086f6fdd0ee80fbd5f75e50fe39" Mar 17 09:31:43 crc kubenswrapper[4813]: I0317 09:31:43.990647 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-mxghx" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.011144 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-27g9l" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.130002 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5c6cd309-922e-4be6-b2d9-12d2aca54b78-db-sync-config-data\") pod \"5c6cd309-922e-4be6-b2d9-12d2aca54b78\" (UID: \"5c6cd309-922e-4be6-b2d9-12d2aca54b78\") " Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.130498 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lgwjb\" (UniqueName: \"kubernetes.io/projected/5c6cd309-922e-4be6-b2d9-12d2aca54b78-kube-api-access-lgwjb\") pod \"5c6cd309-922e-4be6-b2d9-12d2aca54b78\" (UID: \"5c6cd309-922e-4be6-b2d9-12d2aca54b78\") " Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.130692 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41809604-4a36-4453-b381-66b4c5cf1c43-etc-machine-id\") pod \"41809604-4a36-4453-b381-66b4c5cf1c43\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.130798 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/41809604-4a36-4453-b381-66b4c5cf1c43-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "41809604-4a36-4453-b381-66b4c5cf1c43" (UID: "41809604-4a36-4453-b381-66b4c5cf1c43"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.131018 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-db-sync-config-data\") pod \"41809604-4a36-4453-b381-66b4c5cf1c43\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.131172 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zfpnh\" (UniqueName: \"kubernetes.io/projected/41809604-4a36-4453-b381-66b4c5cf1c43-kube-api-access-zfpnh\") pod \"41809604-4a36-4453-b381-66b4c5cf1c43\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.131376 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-config-data\") pod \"41809604-4a36-4453-b381-66b4c5cf1c43\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.131506 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-combined-ca-bundle\") pod \"41809604-4a36-4453-b381-66b4c5cf1c43\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.131719 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-scripts\") pod \"41809604-4a36-4453-b381-66b4c5cf1c43\" (UID: \"41809604-4a36-4453-b381-66b4c5cf1c43\") " Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.131865 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c6cd309-922e-4be6-b2d9-12d2aca54b78-combined-ca-bundle\") pod \"5c6cd309-922e-4be6-b2d9-12d2aca54b78\" (UID: \"5c6cd309-922e-4be6-b2d9-12d2aca54b78\") " Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.132577 4813 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/41809604-4a36-4453-b381-66b4c5cf1c43-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.136439 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41809604-4a36-4453-b381-66b4c5cf1c43-kube-api-access-zfpnh" (OuterVolumeSpecName: "kube-api-access-zfpnh") pod "41809604-4a36-4453-b381-66b4c5cf1c43" (UID: "41809604-4a36-4453-b381-66b4c5cf1c43"). InnerVolumeSpecName "kube-api-access-zfpnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.151340 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "41809604-4a36-4453-b381-66b4c5cf1c43" (UID: "41809604-4a36-4453-b381-66b4c5cf1c43"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.151500 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c6cd309-922e-4be6-b2d9-12d2aca54b78-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "5c6cd309-922e-4be6-b2d9-12d2aca54b78" (UID: "5c6cd309-922e-4be6-b2d9-12d2aca54b78"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.156495 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c6cd309-922e-4be6-b2d9-12d2aca54b78-kube-api-access-lgwjb" (OuterVolumeSpecName: "kube-api-access-lgwjb") pod "5c6cd309-922e-4be6-b2d9-12d2aca54b78" (UID: "5c6cd309-922e-4be6-b2d9-12d2aca54b78"). InnerVolumeSpecName "kube-api-access-lgwjb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.160951 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c6cd309-922e-4be6-b2d9-12d2aca54b78-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c6cd309-922e-4be6-b2d9-12d2aca54b78" (UID: "5c6cd309-922e-4be6-b2d9-12d2aca54b78"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.163746 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-scripts" (OuterVolumeSpecName: "scripts") pod "41809604-4a36-4453-b381-66b4c5cf1c43" (UID: "41809604-4a36-4453-b381-66b4c5cf1c43"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.168989 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41809604-4a36-4453-b381-66b4c5cf1c43" (UID: "41809604-4a36-4453-b381-66b4c5cf1c43"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.187709 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-config-data" (OuterVolumeSpecName: "config-data") pod "41809604-4a36-4453-b381-66b4c5cf1c43" (UID: "41809604-4a36-4453-b381-66b4c5cf1c43"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.239862 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.240186 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.240296 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.240367 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c6cd309-922e-4be6-b2d9-12d2aca54b78-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.240495 4813 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/5c6cd309-922e-4be6-b2d9-12d2aca54b78-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.242005 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lgwjb\" (UniqueName: \"kubernetes.io/projected/5c6cd309-922e-4be6-b2d9-12d2aca54b78-kube-api-access-lgwjb\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.242046 4813 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/41809604-4a36-4453-b381-66b4c5cf1c43-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.242060 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zfpnh\" (UniqueName: \"kubernetes.io/projected/41809604-4a36-4453-b381-66b4c5cf1c43-kube-api-access-zfpnh\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.995759 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-mxghx" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.996152 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d927c167-348f-47f8-9860-b21dac3ca309","Type":"ContainerStarted","Data":"db2d07ad344f758a6f4de8415ae1005b4319f5b36eae2e9f1be1bbb8df4134d3"} Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.996256 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-27g9l" Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.996845 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d927c167-348f-47f8-9860-b21dac3ca309" containerName="ceilometer-central-agent" containerID="cri-o://766a0ef148bc4a1cc73bfe82a1e3d829969f16ee65cb1fd39fd23aac1184be02" gracePeriod=30 Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.997153 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d927c167-348f-47f8-9860-b21dac3ca309" containerName="proxy-httpd" containerID="cri-o://db2d07ad344f758a6f4de8415ae1005b4319f5b36eae2e9f1be1bbb8df4134d3" gracePeriod=30 Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.997264 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d927c167-348f-47f8-9860-b21dac3ca309" containerName="sg-core" containerID="cri-o://faad4b4a7c4535789ba6e076f9588c4385a53196f95b284dcf17a03fd3bc815f" gracePeriod=30 Mar 17 09:31:44 crc kubenswrapper[4813]: I0317 09:31:44.997425 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d927c167-348f-47f8-9860-b21dac3ca309" containerName="ceilometer-notification-agent" containerID="cri-o://996ef8dc81ee850c6bcdf1656dc2e58131b9cf47f3b83174cf9f63eaddb3e604" gracePeriod=30 Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.078881 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.029317614 podStartE2EDuration="55.078851364s" podCreationTimestamp="2026-03-17 09:30:50 +0000 UTC" firstStartedPulling="2026-03-17 09:30:51.637381347 +0000 UTC m=+1273.738184846" lastFinishedPulling="2026-03-17 09:31:44.686915097 +0000 UTC m=+1326.787718596" observedRunningTime="2026-03-17 09:31:45.033680023 +0000 UTC m=+1327.134483522" watchObservedRunningTime="2026-03-17 09:31:45.078851364 +0000 UTC m=+1327.179654863" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.319672 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-5ff5576dbb-hlqck"] Mar 17 09:31:45 crc kubenswrapper[4813]: E0317 09:31:45.320151 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1596452f-363d-4d92-82de-f96f1f83993a" containerName="dnsmasq-dns" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.320170 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1596452f-363d-4d92-82de-f96f1f83993a" containerName="dnsmasq-dns" Mar 17 09:31:45 crc kubenswrapper[4813]: E0317 09:31:45.320200 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41809604-4a36-4453-b381-66b4c5cf1c43" containerName="cinder-db-sync" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.320208 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="41809604-4a36-4453-b381-66b4c5cf1c43" containerName="cinder-db-sync" Mar 17 09:31:45 crc kubenswrapper[4813]: E0317 09:31:45.320224 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1596452f-363d-4d92-82de-f96f1f83993a" containerName="init" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.320232 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1596452f-363d-4d92-82de-f96f1f83993a" containerName="init" Mar 17 09:31:45 crc kubenswrapper[4813]: E0317 09:31:45.320262 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c6cd309-922e-4be6-b2d9-12d2aca54b78" containerName="barbican-db-sync" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.320270 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c6cd309-922e-4be6-b2d9-12d2aca54b78" containerName="barbican-db-sync" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.320459 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="1596452f-363d-4d92-82de-f96f1f83993a" containerName="dnsmasq-dns" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.320477 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="41809604-4a36-4453-b381-66b4c5cf1c43" containerName="cinder-db-sync" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.320487 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c6cd309-922e-4be6-b2d9-12d2aca54b78" containerName="barbican-db-sync" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.321765 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5ff5576dbb-hlqck" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.332008 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-kkzw4" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.332284 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.332352 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.345766 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-76556857b7-95bv6"] Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.351189 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-76556857b7-95bv6" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.368337 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.375369 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5ff5576dbb-hlqck"] Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.411669 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-76556857b7-95bv6"] Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.457657 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.459222 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.469066 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.473750 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f4bba0-7f6d-43cf-8887-00b081856e89-combined-ca-bundle\") pod \"barbican-worker-76556857b7-95bv6\" (UID: \"91f4bba0-7f6d-43cf-8887-00b081856e89\") " pod="openstack/barbican-worker-76556857b7-95bv6" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.473851 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f4bba0-7f6d-43cf-8887-00b081856e89-config-data\") pod \"barbican-worker-76556857b7-95bv6\" (UID: \"91f4bba0-7f6d-43cf-8887-00b081856e89\") " pod="openstack/barbican-worker-76556857b7-95bv6" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.473889 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d44bf48-4229-453c-925d-9653d547de9e-logs\") pod \"barbican-keystone-listener-5ff5576dbb-hlqck\" (UID: \"2d44bf48-4229-453c-925d-9653d547de9e\") " pod="openstack/barbican-keystone-listener-5ff5576dbb-hlqck" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.473935 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d44bf48-4229-453c-925d-9653d547de9e-config-data\") pod \"barbican-keystone-listener-5ff5576dbb-hlqck\" (UID: \"2d44bf48-4229-453c-925d-9653d547de9e\") " pod="openstack/barbican-keystone-listener-5ff5576dbb-hlqck" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.473974 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91f4bba0-7f6d-43cf-8887-00b081856e89-logs\") pod \"barbican-worker-76556857b7-95bv6\" (UID: \"91f4bba0-7f6d-43cf-8887-00b081856e89\") " pod="openstack/barbican-worker-76556857b7-95bv6" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.474004 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5mjv\" (UniqueName: \"kubernetes.io/projected/2d44bf48-4229-453c-925d-9653d547de9e-kube-api-access-r5mjv\") pod \"barbican-keystone-listener-5ff5576dbb-hlqck\" (UID: \"2d44bf48-4229-453c-925d-9653d547de9e\") " pod="openstack/barbican-keystone-listener-5ff5576dbb-hlqck" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.474035 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d44bf48-4229-453c-925d-9653d547de9e-config-data-custom\") pod \"barbican-keystone-listener-5ff5576dbb-hlqck\" (UID: \"2d44bf48-4229-453c-925d-9653d547de9e\") " pod="openstack/barbican-keystone-listener-5ff5576dbb-hlqck" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.474065 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91f4bba0-7f6d-43cf-8887-00b081856e89-config-data-custom\") pod \"barbican-worker-76556857b7-95bv6\" (UID: \"91f4bba0-7f6d-43cf-8887-00b081856e89\") " pod="openstack/barbican-worker-76556857b7-95bv6" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.474108 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nj6ll\" (UniqueName: \"kubernetes.io/projected/91f4bba0-7f6d-43cf-8887-00b081856e89-kube-api-access-nj6ll\") pod \"barbican-worker-76556857b7-95bv6\" (UID: \"91f4bba0-7f6d-43cf-8887-00b081856e89\") " pod="openstack/barbican-worker-76556857b7-95bv6" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.474159 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d44bf48-4229-453c-925d-9653d547de9e-combined-ca-bundle\") pod \"barbican-keystone-listener-5ff5576dbb-hlqck\" (UID: \"2d44bf48-4229-453c-925d-9653d547de9e\") " pod="openstack/barbican-keystone-listener-5ff5576dbb-hlqck" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.480138 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-6rfxv" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.480551 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.480828 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.480985 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.553675 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-h62lf"] Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.555585 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.575814 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d44bf48-4229-453c-925d-9653d547de9e-combined-ca-bundle\") pod \"barbican-keystone-listener-5ff5576dbb-hlqck\" (UID: \"2d44bf48-4229-453c-925d-9653d547de9e\") " pod="openstack/barbican-keystone-listener-5ff5576dbb-hlqck" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.575859 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " pod="openstack/cinder-scheduler-0" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.575894 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " pod="openstack/cinder-scheduler-0" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.575920 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f4bba0-7f6d-43cf-8887-00b081856e89-combined-ca-bundle\") pod \"barbican-worker-76556857b7-95bv6\" (UID: \"91f4bba0-7f6d-43cf-8887-00b081856e89\") " pod="openstack/barbican-worker-76556857b7-95bv6" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.575958 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-scripts\") pod \"cinder-scheduler-0\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " pod="openstack/cinder-scheduler-0" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.576000 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mj4k8\" (UniqueName: \"kubernetes.io/projected/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-kube-api-access-mj4k8\") pod \"cinder-scheduler-0\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " pod="openstack/cinder-scheduler-0" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.576020 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f4bba0-7f6d-43cf-8887-00b081856e89-config-data\") pod \"barbican-worker-76556857b7-95bv6\" (UID: \"91f4bba0-7f6d-43cf-8887-00b081856e89\") " pod="openstack/barbican-worker-76556857b7-95bv6" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.576041 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d44bf48-4229-453c-925d-9653d547de9e-logs\") pod \"barbican-keystone-listener-5ff5576dbb-hlqck\" (UID: \"2d44bf48-4229-453c-925d-9653d547de9e\") " pod="openstack/barbican-keystone-listener-5ff5576dbb-hlqck" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.576073 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d44bf48-4229-453c-925d-9653d547de9e-config-data\") pod \"barbican-keystone-listener-5ff5576dbb-hlqck\" (UID: \"2d44bf48-4229-453c-925d-9653d547de9e\") " pod="openstack/barbican-keystone-listener-5ff5576dbb-hlqck" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.576102 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91f4bba0-7f6d-43cf-8887-00b081856e89-logs\") pod \"barbican-worker-76556857b7-95bv6\" (UID: \"91f4bba0-7f6d-43cf-8887-00b081856e89\") " pod="openstack/barbican-worker-76556857b7-95bv6" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.576123 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5mjv\" (UniqueName: \"kubernetes.io/projected/2d44bf48-4229-453c-925d-9653d547de9e-kube-api-access-r5mjv\") pod \"barbican-keystone-listener-5ff5576dbb-hlqck\" (UID: \"2d44bf48-4229-453c-925d-9653d547de9e\") " pod="openstack/barbican-keystone-listener-5ff5576dbb-hlqck" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.576141 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-config-data\") pod \"cinder-scheduler-0\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " pod="openstack/cinder-scheduler-0" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.576157 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d44bf48-4229-453c-925d-9653d547de9e-config-data-custom\") pod \"barbican-keystone-listener-5ff5576dbb-hlqck\" (UID: \"2d44bf48-4229-453c-925d-9653d547de9e\") " pod="openstack/barbican-keystone-listener-5ff5576dbb-hlqck" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.576174 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91f4bba0-7f6d-43cf-8887-00b081856e89-config-data-custom\") pod \"barbican-worker-76556857b7-95bv6\" (UID: \"91f4bba0-7f6d-43cf-8887-00b081856e89\") " pod="openstack/barbican-worker-76556857b7-95bv6" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.576190 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " pod="openstack/cinder-scheduler-0" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.576217 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nj6ll\" (UniqueName: \"kubernetes.io/projected/91f4bba0-7f6d-43cf-8887-00b081856e89-kube-api-access-nj6ll\") pod \"barbican-worker-76556857b7-95bv6\" (UID: \"91f4bba0-7f6d-43cf-8887-00b081856e89\") " pod="openstack/barbican-worker-76556857b7-95bv6" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.579329 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d44bf48-4229-453c-925d-9653d547de9e-logs\") pod \"barbican-keystone-listener-5ff5576dbb-hlqck\" (UID: \"2d44bf48-4229-453c-925d-9653d547de9e\") " pod="openstack/barbican-keystone-listener-5ff5576dbb-hlqck" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.600530 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/91f4bba0-7f6d-43cf-8887-00b081856e89-logs\") pod \"barbican-worker-76556857b7-95bv6\" (UID: \"91f4bba0-7f6d-43cf-8887-00b081856e89\") " pod="openstack/barbican-worker-76556857b7-95bv6" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.605552 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d44bf48-4229-453c-925d-9653d547de9e-combined-ca-bundle\") pod \"barbican-keystone-listener-5ff5576dbb-hlqck\" (UID: \"2d44bf48-4229-453c-925d-9653d547de9e\") " pod="openstack/barbican-keystone-listener-5ff5576dbb-hlqck" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.606663 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-h62lf"] Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.618252 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nj6ll\" (UniqueName: \"kubernetes.io/projected/91f4bba0-7f6d-43cf-8887-00b081856e89-kube-api-access-nj6ll\") pod \"barbican-worker-76556857b7-95bv6\" (UID: \"91f4bba0-7f6d-43cf-8887-00b081856e89\") " pod="openstack/barbican-worker-76556857b7-95bv6" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.618331 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/91f4bba0-7f6d-43cf-8887-00b081856e89-config-data-custom\") pod \"barbican-worker-76556857b7-95bv6\" (UID: \"91f4bba0-7f6d-43cf-8887-00b081856e89\") " pod="openstack/barbican-worker-76556857b7-95bv6" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.618662 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/91f4bba0-7f6d-43cf-8887-00b081856e89-combined-ca-bundle\") pod \"barbican-worker-76556857b7-95bv6\" (UID: \"91f4bba0-7f6d-43cf-8887-00b081856e89\") " pod="openstack/barbican-worker-76556857b7-95bv6" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.618759 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d44bf48-4229-453c-925d-9653d547de9e-config-data\") pod \"barbican-keystone-listener-5ff5576dbb-hlqck\" (UID: \"2d44bf48-4229-453c-925d-9653d547de9e\") " pod="openstack/barbican-keystone-listener-5ff5576dbb-hlqck" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.623009 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/91f4bba0-7f6d-43cf-8887-00b081856e89-config-data\") pod \"barbican-worker-76556857b7-95bv6\" (UID: \"91f4bba0-7f6d-43cf-8887-00b081856e89\") " pod="openstack/barbican-worker-76556857b7-95bv6" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.631263 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2d44bf48-4229-453c-925d-9653d547de9e-config-data-custom\") pod \"barbican-keystone-listener-5ff5576dbb-hlqck\" (UID: \"2d44bf48-4229-453c-925d-9653d547de9e\") " pod="openstack/barbican-keystone-listener-5ff5576dbb-hlqck" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.644912 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5mjv\" (UniqueName: \"kubernetes.io/projected/2d44bf48-4229-453c-925d-9653d547de9e-kube-api-access-r5mjv\") pod \"barbican-keystone-listener-5ff5576dbb-hlqck\" (UID: \"2d44bf48-4229-453c-925d-9653d547de9e\") " pod="openstack/barbican-keystone-listener-5ff5576dbb-hlqck" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.671584 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-5ff5576dbb-hlqck" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.678515 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-config-data\") pod \"cinder-scheduler-0\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " pod="openstack/cinder-scheduler-0" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.678564 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " pod="openstack/cinder-scheduler-0" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.678585 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-h62lf\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.678679 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " pod="openstack/cinder-scheduler-0" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.678705 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " pod="openstack/cinder-scheduler-0" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.678732 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-config\") pod \"dnsmasq-dns-75c8ddd69c-h62lf\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.678747 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-scripts\") pod \"cinder-scheduler-0\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " pod="openstack/cinder-scheduler-0" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.678783 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mj4k8\" (UniqueName: \"kubernetes.io/projected/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-kube-api-access-mj4k8\") pod \"cinder-scheduler-0\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " pod="openstack/cinder-scheduler-0" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.678806 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-h62lf\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.678830 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-h62lf\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.678853 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-h62lf\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.678866 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-znz64\" (UniqueName: \"kubernetes.io/projected/d0383aaf-be20-446a-97fe-c7d959a3b9f1-kube-api-access-znz64\") pod \"dnsmasq-dns-75c8ddd69c-h62lf\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.689539 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " pod="openstack/cinder-scheduler-0" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.689668 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " pod="openstack/cinder-scheduler-0" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.707713 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-h62lf"] Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.711236 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-76556857b7-95bv6" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.720552 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-config-data\") pod \"cinder-scheduler-0\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " pod="openstack/cinder-scheduler-0" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.720661 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-scripts\") pod \"cinder-scheduler-0\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " pod="openstack/cinder-scheduler-0" Mar 17 09:31:45 crc kubenswrapper[4813]: E0317 09:31:45.721394 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-znz64 ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" podUID="d0383aaf-be20-446a-97fe-c7d959a3b9f1" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.730568 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " pod="openstack/cinder-scheduler-0" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.739008 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mj4k8\" (UniqueName: \"kubernetes.io/projected/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-kube-api-access-mj4k8\") pod \"cinder-scheduler-0\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " pod="openstack/cinder-scheduler-0" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.765775 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-69677499d6-qsdxd"] Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.767251 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.769726 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.793821 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-h62lf\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.793990 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-config\") pod \"dnsmasq-dns-75c8ddd69c-h62lf\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.794061 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-h62lf\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.794085 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-h62lf\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.794111 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-h62lf\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.794129 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-znz64\" (UniqueName: \"kubernetes.io/projected/d0383aaf-be20-446a-97fe-c7d959a3b9f1-kube-api-access-znz64\") pod \"dnsmasq-dns-75c8ddd69c-h62lf\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.795294 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-dns-svc\") pod \"dnsmasq-dns-75c8ddd69c-h62lf\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.795336 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-dns-swift-storage-0\") pod \"dnsmasq-dns-75c8ddd69c-h62lf\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.800091 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-69677499d6-qsdxd"] Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.817232 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.820165 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-ovsdbserver-sb\") pod \"dnsmasq-dns-75c8ddd69c-h62lf\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.820413 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-config\") pod \"dnsmasq-dns-75c8ddd69c-h62lf\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.821938 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-ovsdbserver-nb\") pod \"dnsmasq-dns-75c8ddd69c-h62lf\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.858509 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-znz64\" (UniqueName: \"kubernetes.io/projected/d0383aaf-be20-446a-97fe-c7d959a3b9f1-kube-api-access-znz64\") pod \"dnsmasq-dns-75c8ddd69c-h62lf\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.893129 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-7wdg7"] Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.897575 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.903947 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z448k\" (UniqueName: \"kubernetes.io/projected/e1545875-bc39-40cd-a76a-6229690912a2-kube-api-access-z448k\") pod \"barbican-api-69677499d6-qsdxd\" (UID: \"e1545875-bc39-40cd-a76a-6229690912a2\") " pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.904022 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1545875-bc39-40cd-a76a-6229690912a2-combined-ca-bundle\") pod \"barbican-api-69677499d6-qsdxd\" (UID: \"e1545875-bc39-40cd-a76a-6229690912a2\") " pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.904113 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e1545875-bc39-40cd-a76a-6229690912a2-config-data-custom\") pod \"barbican-api-69677499d6-qsdxd\" (UID: \"e1545875-bc39-40cd-a76a-6229690912a2\") " pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.904132 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1545875-bc39-40cd-a76a-6229690912a2-logs\") pod \"barbican-api-69677499d6-qsdxd\" (UID: \"e1545875-bc39-40cd-a76a-6229690912a2\") " pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.904161 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1545875-bc39-40cd-a76a-6229690912a2-config-data\") pod \"barbican-api-69677499d6-qsdxd\" (UID: \"e1545875-bc39-40cd-a76a-6229690912a2\") " pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.911208 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-7wdg7"] Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.934089 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.935490 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.942018 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 17 09:31:45 crc kubenswrapper[4813]: I0317 09:31:45.965703 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.006070 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-7wdg7\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.006112 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-config-data-custom\") pod \"cinder-api-0\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.006143 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.006173 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e1545875-bc39-40cd-a76a-6229690912a2-config-data-custom\") pod \"barbican-api-69677499d6-qsdxd\" (UID: \"e1545875-bc39-40cd-a76a-6229690912a2\") " pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.006191 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1545875-bc39-40cd-a76a-6229690912a2-logs\") pod \"barbican-api-69677499d6-qsdxd\" (UID: \"e1545875-bc39-40cd-a76a-6229690912a2\") " pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.006211 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96lv2\" (UniqueName: \"kubernetes.io/projected/ae0c0042-8dd1-44dc-a09f-5fefa2106098-kube-api-access-96lv2\") pod \"cinder-api-0\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.006258 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1545875-bc39-40cd-a76a-6229690912a2-config-data\") pod \"barbican-api-69677499d6-qsdxd\" (UID: \"e1545875-bc39-40cd-a76a-6229690912a2\") " pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.006286 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-7wdg7\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.006314 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-scripts\") pod \"cinder-api-0\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.006332 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-config\") pod \"dnsmasq-dns-5784cf869f-7wdg7\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.006368 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-7wdg7\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.006389 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnrjq\" (UniqueName: \"kubernetes.io/projected/899c033d-3f0f-450b-8c73-227886dc17e0-kube-api-access-fnrjq\") pod \"dnsmasq-dns-5784cf869f-7wdg7\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.006407 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-dns-svc\") pod \"dnsmasq-dns-5784cf869f-7wdg7\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.006445 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae0c0042-8dd1-44dc-a09f-5fefa2106098-logs\") pod \"cinder-api-0\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.006463 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-config-data\") pod \"cinder-api-0\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.006479 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae0c0042-8dd1-44dc-a09f-5fefa2106098-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.006511 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z448k\" (UniqueName: \"kubernetes.io/projected/e1545875-bc39-40cd-a76a-6229690912a2-kube-api-access-z448k\") pod \"barbican-api-69677499d6-qsdxd\" (UID: \"e1545875-bc39-40cd-a76a-6229690912a2\") " pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.006557 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1545875-bc39-40cd-a76a-6229690912a2-combined-ca-bundle\") pod \"barbican-api-69677499d6-qsdxd\" (UID: \"e1545875-bc39-40cd-a76a-6229690912a2\") " pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.009496 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1545875-bc39-40cd-a76a-6229690912a2-logs\") pod \"barbican-api-69677499d6-qsdxd\" (UID: \"e1545875-bc39-40cd-a76a-6229690912a2\") " pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.017075 4813 generic.go:334] "Generic (PLEG): container finished" podID="d927c167-348f-47f8-9860-b21dac3ca309" containerID="faad4b4a7c4535789ba6e076f9588c4385a53196f95b284dcf17a03fd3bc815f" exitCode=2 Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.017155 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.017812 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d927c167-348f-47f8-9860-b21dac3ca309","Type":"ContainerDied","Data":"faad4b4a7c4535789ba6e076f9588c4385a53196f95b284dcf17a03fd3bc815f"} Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.025382 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1545875-bc39-40cd-a76a-6229690912a2-combined-ca-bundle\") pod \"barbican-api-69677499d6-qsdxd\" (UID: \"e1545875-bc39-40cd-a76a-6229690912a2\") " pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.027582 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e1545875-bc39-40cd-a76a-6229690912a2-config-data-custom\") pod \"barbican-api-69677499d6-qsdxd\" (UID: \"e1545875-bc39-40cd-a76a-6229690912a2\") " pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.029195 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z448k\" (UniqueName: \"kubernetes.io/projected/e1545875-bc39-40cd-a76a-6229690912a2-kube-api-access-z448k\") pod \"barbican-api-69677499d6-qsdxd\" (UID: \"e1545875-bc39-40cd-a76a-6229690912a2\") " pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.033046 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1545875-bc39-40cd-a76a-6229690912a2-config-data\") pod \"barbican-api-69677499d6-qsdxd\" (UID: \"e1545875-bc39-40cd-a76a-6229690912a2\") " pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.108554 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-7wdg7\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.108839 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-config-data-custom\") pod \"cinder-api-0\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.108892 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.108919 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96lv2\" (UniqueName: \"kubernetes.io/projected/ae0c0042-8dd1-44dc-a09f-5fefa2106098-kube-api-access-96lv2\") pod \"cinder-api-0\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.108972 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-7wdg7\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.108999 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-scripts\") pod \"cinder-api-0\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.109016 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-config\") pod \"dnsmasq-dns-5784cf869f-7wdg7\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.109051 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-7wdg7\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.109073 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnrjq\" (UniqueName: \"kubernetes.io/projected/899c033d-3f0f-450b-8c73-227886dc17e0-kube-api-access-fnrjq\") pod \"dnsmasq-dns-5784cf869f-7wdg7\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.109090 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-dns-svc\") pod \"dnsmasq-dns-5784cf869f-7wdg7\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.109112 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae0c0042-8dd1-44dc-a09f-5fefa2106098-logs\") pod \"cinder-api-0\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.109129 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-config-data\") pod \"cinder-api-0\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.109145 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae0c0042-8dd1-44dc-a09f-5fefa2106098-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.109233 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae0c0042-8dd1-44dc-a09f-5fefa2106098-etc-machine-id\") pod \"cinder-api-0\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.110180 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-dns-swift-storage-0\") pod \"dnsmasq-dns-5784cf869f-7wdg7\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.110321 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-dns-svc\") pod \"dnsmasq-dns-5784cf869f-7wdg7\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.110455 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae0c0042-8dd1-44dc-a09f-5fefa2106098-logs\") pod \"cinder-api-0\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.110583 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-config\") pod \"dnsmasq-dns-5784cf869f-7wdg7\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.113756 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-ovsdbserver-nb\") pod \"dnsmasq-dns-5784cf869f-7wdg7\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.114333 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-ovsdbserver-sb\") pod \"dnsmasq-dns-5784cf869f-7wdg7\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.118664 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-config-data\") pod \"cinder-api-0\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.129626 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.137240 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-scripts\") pod \"cinder-api-0\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.137277 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.149882 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-config-data-custom\") pod \"cinder-api-0\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.159928 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnrjq\" (UniqueName: \"kubernetes.io/projected/899c033d-3f0f-450b-8c73-227886dc17e0-kube-api-access-fnrjq\") pod \"dnsmasq-dns-5784cf869f-7wdg7\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.161067 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96lv2\" (UniqueName: \"kubernetes.io/projected/ae0c0042-8dd1-44dc-a09f-5fefa2106098-kube-api-access-96lv2\") pod \"cinder-api-0\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.265686 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.275500 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.310697 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.415246 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-znz64\" (UniqueName: \"kubernetes.io/projected/d0383aaf-be20-446a-97fe-c7d959a3b9f1-kube-api-access-znz64\") pod \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.415301 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-ovsdbserver-nb\") pod \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.415332 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-dns-svc\") pod \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.415469 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-dns-swift-storage-0\") pod \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.415585 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-ovsdbserver-sb\") pod \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.415674 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-config\") pod \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\" (UID: \"d0383aaf-be20-446a-97fe-c7d959a3b9f1\") " Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.418262 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d0383aaf-be20-446a-97fe-c7d959a3b9f1" (UID: "d0383aaf-be20-446a-97fe-c7d959a3b9f1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.418566 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d0383aaf-be20-446a-97fe-c7d959a3b9f1" (UID: "d0383aaf-be20-446a-97fe-c7d959a3b9f1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.418865 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "d0383aaf-be20-446a-97fe-c7d959a3b9f1" (UID: "d0383aaf-be20-446a-97fe-c7d959a3b9f1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.419149 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d0383aaf-be20-446a-97fe-c7d959a3b9f1" (UID: "d0383aaf-be20-446a-97fe-c7d959a3b9f1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.419434 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-config" (OuterVolumeSpecName: "config") pod "d0383aaf-be20-446a-97fe-c7d959a3b9f1" (UID: "d0383aaf-be20-446a-97fe-c7d959a3b9f1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.428226 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0383aaf-be20-446a-97fe-c7d959a3b9f1-kube-api-access-znz64" (OuterVolumeSpecName: "kube-api-access-znz64") pod "d0383aaf-be20-446a-97fe-c7d959a3b9f1" (UID: "d0383aaf-be20-446a-97fe-c7d959a3b9f1"). InnerVolumeSpecName "kube-api-access-znz64". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.519740 4813 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.519778 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.519789 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.519798 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-znz64\" (UniqueName: \"kubernetes.io/projected/d0383aaf-be20-446a-97fe-c7d959a3b9f1-kube-api-access-znz64\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.519812 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.519821 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0383aaf-be20-446a-97fe-c7d959a3b9f1-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.643282 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-76556857b7-95bv6"] Mar 17 09:31:46 crc kubenswrapper[4813]: W0317 09:31:46.671888 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2d44bf48_4229_453c_925d_9653d547de9e.slice/crio-c12ed051ec928990c063d4ea944bc40f2108eda1040c6640cdbf3336bb3685ae WatchSource:0}: Error finding container c12ed051ec928990c063d4ea944bc40f2108eda1040c6640cdbf3336bb3685ae: Status 404 returned error can't find the container with id c12ed051ec928990c063d4ea944bc40f2108eda1040c6640cdbf3336bb3685ae Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.678108 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-5ff5576dbb-hlqck"] Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.778697 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 17 09:31:46 crc kubenswrapper[4813]: W0317 09:31:46.894266 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1545875_bc39_40cd_a76a_6229690912a2.slice/crio-5429603a853162372b069be1edffa26fc59fd14f2b8b7a6e6c6401e5f124505b WatchSource:0}: Error finding container 5429603a853162372b069be1edffa26fc59fd14f2b8b7a6e6c6401e5f124505b: Status 404 returned error can't find the container with id 5429603a853162372b069be1edffa26fc59fd14f2b8b7a6e6c6401e5f124505b Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.894815 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-69677499d6-qsdxd"] Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.920204 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-7wdg7"] Mar 17 09:31:46 crc kubenswrapper[4813]: I0317 09:31:46.930319 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 17 09:31:47 crc kubenswrapper[4813]: I0317 09:31:47.049557 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76556857b7-95bv6" event={"ID":"91f4bba0-7f6d-43cf-8887-00b081856e89","Type":"ContainerStarted","Data":"471a8125e1c6a02bc882a2bfd8f11844be7ac2a35b02950cd8d5af5c6f709256"} Mar 17 09:31:47 crc kubenswrapper[4813]: I0317 09:31:47.056033 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" event={"ID":"899c033d-3f0f-450b-8c73-227886dc17e0","Type":"ContainerStarted","Data":"aeb0257f37af242b6faeda2ee02ffd22232fe3a8216b2dcaa2b1a8a9387c0faf"} Mar 17 09:31:47 crc kubenswrapper[4813]: I0317 09:31:47.059776 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-69677499d6-qsdxd" event={"ID":"e1545875-bc39-40cd-a76a-6229690912a2","Type":"ContainerStarted","Data":"5429603a853162372b069be1edffa26fc59fd14f2b8b7a6e6c6401e5f124505b"} Mar 17 09:31:47 crc kubenswrapper[4813]: I0317 09:31:47.063281 4813 generic.go:334] "Generic (PLEG): container finished" podID="d927c167-348f-47f8-9860-b21dac3ca309" containerID="766a0ef148bc4a1cc73bfe82a1e3d829969f16ee65cb1fd39fd23aac1184be02" exitCode=0 Mar 17 09:31:47 crc kubenswrapper[4813]: I0317 09:31:47.063326 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d927c167-348f-47f8-9860-b21dac3ca309","Type":"ContainerDied","Data":"766a0ef148bc4a1cc73bfe82a1e3d829969f16ee65cb1fd39fd23aac1184be02"} Mar 17 09:31:47 crc kubenswrapper[4813]: I0317 09:31:47.066062 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3a5534d6-3502-4a49-8991-cf7fe05d2e8d","Type":"ContainerStarted","Data":"79a39c8324945965793609c773665e493204f54fe03d7d4c9b01ebb2a9f32165"} Mar 17 09:31:47 crc kubenswrapper[4813]: I0317 09:31:47.067692 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ae0c0042-8dd1-44dc-a09f-5fefa2106098","Type":"ContainerStarted","Data":"181b0cddfb2dcf530c43551e662f361f387f4d7105aa65c1bf9607b0974f8c6b"} Mar 17 09:31:47 crc kubenswrapper[4813]: I0317 09:31:47.072016 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75c8ddd69c-h62lf" Mar 17 09:31:47 crc kubenswrapper[4813]: I0317 09:31:47.072895 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5ff5576dbb-hlqck" event={"ID":"2d44bf48-4229-453c-925d-9653d547de9e","Type":"ContainerStarted","Data":"c12ed051ec928990c063d4ea944bc40f2108eda1040c6640cdbf3336bb3685ae"} Mar 17 09:31:47 crc kubenswrapper[4813]: I0317 09:31:47.161855 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-h62lf"] Mar 17 09:31:47 crc kubenswrapper[4813]: I0317 09:31:47.161909 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75c8ddd69c-h62lf"] Mar 17 09:31:47 crc kubenswrapper[4813]: I0317 09:31:47.611184 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 17 09:31:48 crc kubenswrapper[4813]: I0317 09:31:48.099546 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ae0c0042-8dd1-44dc-a09f-5fefa2106098","Type":"ContainerStarted","Data":"c61ee99e3a934dfa64ff56ca74205da682067f61605c8557823343a06f45d22d"} Mar 17 09:31:48 crc kubenswrapper[4813]: I0317 09:31:48.101823 4813 generic.go:334] "Generic (PLEG): container finished" podID="899c033d-3f0f-450b-8c73-227886dc17e0" containerID="07fc41d3607cf1234be698707cf2db9ace33c8187811d985db0d032100082ac8" exitCode=0 Mar 17 09:31:48 crc kubenswrapper[4813]: I0317 09:31:48.101865 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" event={"ID":"899c033d-3f0f-450b-8c73-227886dc17e0","Type":"ContainerDied","Data":"07fc41d3607cf1234be698707cf2db9ace33c8187811d985db0d032100082ac8"} Mar 17 09:31:48 crc kubenswrapper[4813]: I0317 09:31:48.105354 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-69677499d6-qsdxd" event={"ID":"e1545875-bc39-40cd-a76a-6229690912a2","Type":"ContainerStarted","Data":"fd344086822719f61ad20d2b0c79d5707af9721b8d69451ce2aa6081ba82ae43"} Mar 17 09:31:48 crc kubenswrapper[4813]: I0317 09:31:48.105385 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-69677499d6-qsdxd" event={"ID":"e1545875-bc39-40cd-a76a-6229690912a2","Type":"ContainerStarted","Data":"b0dd7b1cea28102ec7d70d9ec0f5ed7a01e2a0ffe4c718f45dde21ccc0a0cf4e"} Mar 17 09:31:48 crc kubenswrapper[4813]: I0317 09:31:48.105884 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:31:48 crc kubenswrapper[4813]: I0317 09:31:48.105904 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:31:48 crc kubenswrapper[4813]: I0317 09:31:48.108758 4813 generic.go:334] "Generic (PLEG): container finished" podID="d927c167-348f-47f8-9860-b21dac3ca309" containerID="996ef8dc81ee850c6bcdf1656dc2e58131b9cf47f3b83174cf9f63eaddb3e604" exitCode=0 Mar 17 09:31:48 crc kubenswrapper[4813]: I0317 09:31:48.108781 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d927c167-348f-47f8-9860-b21dac3ca309","Type":"ContainerDied","Data":"996ef8dc81ee850c6bcdf1656dc2e58131b9cf47f3b83174cf9f63eaddb3e604"} Mar 17 09:31:48 crc kubenswrapper[4813]: I0317 09:31:48.184810 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-69677499d6-qsdxd" podStartSLOduration=3.184789508 podStartE2EDuration="3.184789508s" podCreationTimestamp="2026-03-17 09:31:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:31:48.183841858 +0000 UTC m=+1330.284645357" watchObservedRunningTime="2026-03-17 09:31:48.184789508 +0000 UTC m=+1330.285593007" Mar 17 09:31:48 crc kubenswrapper[4813]: I0317 09:31:48.784501 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0383aaf-be20-446a-97fe-c7d959a3b9f1" path="/var/lib/kubelet/pods/d0383aaf-be20-446a-97fe-c7d959a3b9f1/volumes" Mar 17 09:31:49 crc kubenswrapper[4813]: I0317 09:31:49.125648 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5ff5576dbb-hlqck" event={"ID":"2d44bf48-4229-453c-925d-9653d547de9e","Type":"ContainerStarted","Data":"fde4cce78253c9ca28cf4929fbdcdce3fb04c71b1047056aa9520d9029387295"} Mar 17 09:31:49 crc kubenswrapper[4813]: I0317 09:31:49.131401 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" event={"ID":"899c033d-3f0f-450b-8c73-227886dc17e0","Type":"ContainerStarted","Data":"4d02f3206c8ca2e0e95ae4f6269f84d8d04299fb1615e599a5f7a19d1ec7a1ad"} Mar 17 09:31:49 crc kubenswrapper[4813]: I0317 09:31:49.131533 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:49 crc kubenswrapper[4813]: I0317 09:31:49.153289 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" podStartSLOduration=4.153269757 podStartE2EDuration="4.153269757s" podCreationTimestamp="2026-03-17 09:31:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:31:49.152882926 +0000 UTC m=+1331.253686425" watchObservedRunningTime="2026-03-17 09:31:49.153269757 +0000 UTC m=+1331.254073266" Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.142649 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76556857b7-95bv6" event={"ID":"91f4bba0-7f6d-43cf-8887-00b081856e89","Type":"ContainerStarted","Data":"47ee2750cb588adb9c3a8700caa9d6353e2f96930d515b18c4e82b62908a7777"} Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.143001 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-76556857b7-95bv6" event={"ID":"91f4bba0-7f6d-43cf-8887-00b081856e89","Type":"ContainerStarted","Data":"6a4cd270054091303f45779f623aa83efd8a08f1035fdb809e346f1a21be05cb"} Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.145518 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3a5534d6-3502-4a49-8991-cf7fe05d2e8d","Type":"ContainerStarted","Data":"50b311fe853f38caed37e1ab6162fade6400373689947e62efdaea2a928d8569"} Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.145957 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3a5534d6-3502-4a49-8991-cf7fe05d2e8d","Type":"ContainerStarted","Data":"f1396ad8c478422139fb9b0205ca8715ff597cfab4f6cd9e69e2aa70bb114796"} Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.148321 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ae0c0042-8dd1-44dc-a09f-5fefa2106098","Type":"ContainerStarted","Data":"67365eb2378a9e978597534b8c4765f87835687066c9e69f2fa89376472fdc08"} Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.148417 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ae0c0042-8dd1-44dc-a09f-5fefa2106098" containerName="cinder-api-log" containerID="cri-o://c61ee99e3a934dfa64ff56ca74205da682067f61605c8557823343a06f45d22d" gracePeriod=30 Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.148448 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.148459 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="ae0c0042-8dd1-44dc-a09f-5fefa2106098" containerName="cinder-api" containerID="cri-o://67365eb2378a9e978597534b8c4765f87835687066c9e69f2fa89376472fdc08" gracePeriod=30 Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.151094 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-5ff5576dbb-hlqck" event={"ID":"2d44bf48-4229-453c-925d-9653d547de9e","Type":"ContainerStarted","Data":"ac5183a42365aed2b74f44426332621da5cfe291eeefc0dcedfa6807f5c7fc62"} Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.183911 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-76556857b7-95bv6" podStartSLOduration=3.139165492 podStartE2EDuration="5.183893381s" podCreationTimestamp="2026-03-17 09:31:45 +0000 UTC" firstStartedPulling="2026-03-17 09:31:46.645798035 +0000 UTC m=+1328.746601524" lastFinishedPulling="2026-03-17 09:31:48.690525914 +0000 UTC m=+1330.791329413" observedRunningTime="2026-03-17 09:31:50.179203714 +0000 UTC m=+1332.280007213" watchObservedRunningTime="2026-03-17 09:31:50.183893381 +0000 UTC m=+1332.284696880" Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.212555 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.306917928 podStartE2EDuration="5.212536682s" podCreationTimestamp="2026-03-17 09:31:45 +0000 UTC" firstStartedPulling="2026-03-17 09:31:46.787788381 +0000 UTC m=+1328.888591890" lastFinishedPulling="2026-03-17 09:31:48.693407145 +0000 UTC m=+1330.794210644" observedRunningTime="2026-03-17 09:31:50.201486604 +0000 UTC m=+1332.302290103" watchObservedRunningTime="2026-03-17 09:31:50.212536682 +0000 UTC m=+1332.313340171" Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.228725 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.22870491 podStartE2EDuration="5.22870491s" podCreationTimestamp="2026-03-17 09:31:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:31:50.227004987 +0000 UTC m=+1332.327808486" watchObservedRunningTime="2026-03-17 09:31:50.22870491 +0000 UTC m=+1332.329508409" Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.264678 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-5ff5576dbb-hlqck" podStartSLOduration=3.251839126 podStartE2EDuration="5.264655091s" podCreationTimestamp="2026-03-17 09:31:45 +0000 UTC" firstStartedPulling="2026-03-17 09:31:46.676265313 +0000 UTC m=+1328.777068812" lastFinishedPulling="2026-03-17 09:31:48.689081278 +0000 UTC m=+1330.789884777" observedRunningTime="2026-03-17 09:31:50.24587939 +0000 UTC m=+1332.346682900" watchObservedRunningTime="2026-03-17 09:31:50.264655091 +0000 UTC m=+1332.365458600" Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.510829 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.766684 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.820647 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.821873 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-config-data\") pod \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.821943 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae0c0042-8dd1-44dc-a09f-5fefa2106098-etc-machine-id\") pod \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.822003 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-scripts\") pod \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.822022 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae0c0042-8dd1-44dc-a09f-5fefa2106098-logs\") pod \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.822050 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-config-data-custom\") pod \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.824406 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ae0c0042-8dd1-44dc-a09f-5fefa2106098-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ae0c0042-8dd1-44dc-a09f-5fefa2106098" (UID: "ae0c0042-8dd1-44dc-a09f-5fefa2106098"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.824888 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ae0c0042-8dd1-44dc-a09f-5fefa2106098-logs" (OuterVolumeSpecName: "logs") pod "ae0c0042-8dd1-44dc-a09f-5fefa2106098" (UID: "ae0c0042-8dd1-44dc-a09f-5fefa2106098"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.828772 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "ae0c0042-8dd1-44dc-a09f-5fefa2106098" (UID: "ae0c0042-8dd1-44dc-a09f-5fefa2106098"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.834712 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-scripts" (OuterVolumeSpecName: "scripts") pod "ae0c0042-8dd1-44dc-a09f-5fefa2106098" (UID: "ae0c0042-8dd1-44dc-a09f-5fefa2106098"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.884820 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-config-data" (OuterVolumeSpecName: "config-data") pod "ae0c0042-8dd1-44dc-a09f-5fefa2106098" (UID: "ae0c0042-8dd1-44dc-a09f-5fefa2106098"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.923837 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-combined-ca-bundle\") pod \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.923933 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96lv2\" (UniqueName: \"kubernetes.io/projected/ae0c0042-8dd1-44dc-a09f-5fefa2106098-kube-api-access-96lv2\") pod \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\" (UID: \"ae0c0042-8dd1-44dc-a09f-5fefa2106098\") " Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.924462 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.924476 4813 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ae0c0042-8dd1-44dc-a09f-5fefa2106098-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.924485 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.924495 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ae0c0042-8dd1-44dc-a09f-5fefa2106098-logs\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.924503 4813 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.927931 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae0c0042-8dd1-44dc-a09f-5fefa2106098-kube-api-access-96lv2" (OuterVolumeSpecName: "kube-api-access-96lv2") pod "ae0c0042-8dd1-44dc-a09f-5fefa2106098" (UID: "ae0c0042-8dd1-44dc-a09f-5fefa2106098"). InnerVolumeSpecName "kube-api-access-96lv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:31:50 crc kubenswrapper[4813]: I0317 09:31:50.955067 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae0c0042-8dd1-44dc-a09f-5fefa2106098" (UID: "ae0c0042-8dd1-44dc-a09f-5fefa2106098"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.014755 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-b5495b5bd-g77kp" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.025980 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96lv2\" (UniqueName: \"kubernetes.io/projected/ae0c0042-8dd1-44dc-a09f-5fefa2106098-kube-api-access-96lv2\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.026019 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae0c0042-8dd1-44dc-a09f-5fefa2106098-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.166585 4813 generic.go:334] "Generic (PLEG): container finished" podID="ae0c0042-8dd1-44dc-a09f-5fefa2106098" containerID="67365eb2378a9e978597534b8c4765f87835687066c9e69f2fa89376472fdc08" exitCode=0 Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.166637 4813 generic.go:334] "Generic (PLEG): container finished" podID="ae0c0042-8dd1-44dc-a09f-5fefa2106098" containerID="c61ee99e3a934dfa64ff56ca74205da682067f61605c8557823343a06f45d22d" exitCode=143 Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.166764 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ae0c0042-8dd1-44dc-a09f-5fefa2106098","Type":"ContainerDied","Data":"67365eb2378a9e978597534b8c4765f87835687066c9e69f2fa89376472fdc08"} Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.166803 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ae0c0042-8dd1-44dc-a09f-5fefa2106098","Type":"ContainerDied","Data":"c61ee99e3a934dfa64ff56ca74205da682067f61605c8557823343a06f45d22d"} Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.166843 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"ae0c0042-8dd1-44dc-a09f-5fefa2106098","Type":"ContainerDied","Data":"181b0cddfb2dcf530c43551e662f361f387f4d7105aa65c1bf9607b0974f8c6b"} Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.166860 4813 scope.go:117] "RemoveContainer" containerID="67365eb2378a9e978597534b8c4765f87835687066c9e69f2fa89376472fdc08" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.167007 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.219099 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.219579 4813 scope.go:117] "RemoveContainer" containerID="c61ee99e3a934dfa64ff56ca74205da682067f61605c8557823343a06f45d22d" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.252790 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.267516 4813 scope.go:117] "RemoveContainer" containerID="67365eb2378a9e978597534b8c4765f87835687066c9e69f2fa89376472fdc08" Mar 17 09:31:51 crc kubenswrapper[4813]: E0317 09:31:51.268028 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67365eb2378a9e978597534b8c4765f87835687066c9e69f2fa89376472fdc08\": container with ID starting with 67365eb2378a9e978597534b8c4765f87835687066c9e69f2fa89376472fdc08 not found: ID does not exist" containerID="67365eb2378a9e978597534b8c4765f87835687066c9e69f2fa89376472fdc08" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.268064 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67365eb2378a9e978597534b8c4765f87835687066c9e69f2fa89376472fdc08"} err="failed to get container status \"67365eb2378a9e978597534b8c4765f87835687066c9e69f2fa89376472fdc08\": rpc error: code = NotFound desc = could not find container \"67365eb2378a9e978597534b8c4765f87835687066c9e69f2fa89376472fdc08\": container with ID starting with 67365eb2378a9e978597534b8c4765f87835687066c9e69f2fa89376472fdc08 not found: ID does not exist" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.268089 4813 scope.go:117] "RemoveContainer" containerID="c61ee99e3a934dfa64ff56ca74205da682067f61605c8557823343a06f45d22d" Mar 17 09:31:51 crc kubenswrapper[4813]: E0317 09:31:51.268401 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c61ee99e3a934dfa64ff56ca74205da682067f61605c8557823343a06f45d22d\": container with ID starting with c61ee99e3a934dfa64ff56ca74205da682067f61605c8557823343a06f45d22d not found: ID does not exist" containerID="c61ee99e3a934dfa64ff56ca74205da682067f61605c8557823343a06f45d22d" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.268416 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c61ee99e3a934dfa64ff56ca74205da682067f61605c8557823343a06f45d22d"} err="failed to get container status \"c61ee99e3a934dfa64ff56ca74205da682067f61605c8557823343a06f45d22d\": rpc error: code = NotFound desc = could not find container \"c61ee99e3a934dfa64ff56ca74205da682067f61605c8557823343a06f45d22d\": container with ID starting with c61ee99e3a934dfa64ff56ca74205da682067f61605c8557823343a06f45d22d not found: ID does not exist" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.268428 4813 scope.go:117] "RemoveContainer" containerID="67365eb2378a9e978597534b8c4765f87835687066c9e69f2fa89376472fdc08" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.268588 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67365eb2378a9e978597534b8c4765f87835687066c9e69f2fa89376472fdc08"} err="failed to get container status \"67365eb2378a9e978597534b8c4765f87835687066c9e69f2fa89376472fdc08\": rpc error: code = NotFound desc = could not find container \"67365eb2378a9e978597534b8c4765f87835687066c9e69f2fa89376472fdc08\": container with ID starting with 67365eb2378a9e978597534b8c4765f87835687066c9e69f2fa89376472fdc08 not found: ID does not exist" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.268618 4813 scope.go:117] "RemoveContainer" containerID="c61ee99e3a934dfa64ff56ca74205da682067f61605c8557823343a06f45d22d" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.268775 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c61ee99e3a934dfa64ff56ca74205da682067f61605c8557823343a06f45d22d"} err="failed to get container status \"c61ee99e3a934dfa64ff56ca74205da682067f61605c8557823343a06f45d22d\": rpc error: code = NotFound desc = could not find container \"c61ee99e3a934dfa64ff56ca74205da682067f61605c8557823343a06f45d22d\": container with ID starting with c61ee99e3a934dfa64ff56ca74205da682067f61605c8557823343a06f45d22d not found: ID does not exist" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.275992 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Mar 17 09:31:51 crc kubenswrapper[4813]: E0317 09:31:51.276388 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae0c0042-8dd1-44dc-a09f-5fefa2106098" containerName="cinder-api" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.276402 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae0c0042-8dd1-44dc-a09f-5fefa2106098" containerName="cinder-api" Mar 17 09:31:51 crc kubenswrapper[4813]: E0317 09:31:51.276424 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae0c0042-8dd1-44dc-a09f-5fefa2106098" containerName="cinder-api-log" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.276430 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae0c0042-8dd1-44dc-a09f-5fefa2106098" containerName="cinder-api-log" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.276672 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae0c0042-8dd1-44dc-a09f-5fefa2106098" containerName="cinder-api-log" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.276688 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae0c0042-8dd1-44dc-a09f-5fefa2106098" containerName="cinder-api" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.278191 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.286072 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.286136 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.286178 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.310266 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.344701 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7d558687d7-q6tr2"] Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.344939 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7d558687d7-q6tr2" podUID="5aaee075-cdd4-4532-a651-539dfc8b6ac9" containerName="neutron-api" containerID="cri-o://8319e5f3fa73599000d2c08ec05ceacf5dcc9f68e9da36049d345fa482790c15" gracePeriod=30 Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.345708 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-7d558687d7-q6tr2" podUID="5aaee075-cdd4-4532-a651-539dfc8b6ac9" containerName="neutron-httpd" containerID="cri-o://f4c6cd7b65fba4bb7d14a208ed65138752b3610ba036e75b33bc344a062202be" gracePeriod=30 Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.358391 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-76bd9cc6f7-x2p6w"] Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.359912 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.376563 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-76bd9cc6f7-x2p6w"] Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.432663 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27d30ca5-0f9f-4f22-83a6-66c9e0942930-logs\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.432732 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27d30ca5-0f9f-4f22-83a6-66c9e0942930-config-data\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.432809 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d30ca5-0f9f-4f22-83a6-66c9e0942930-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.432845 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27d30ca5-0f9f-4f22-83a6-66c9e0942930-config-data-custom\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.432945 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27d30ca5-0f9f-4f22-83a6-66c9e0942930-scripts\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.432997 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d30ca5-0f9f-4f22-83a6-66c9e0942930-public-tls-certs\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.433056 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/27d30ca5-0f9f-4f22-83a6-66c9e0942930-etc-machine-id\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.433111 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d30ca5-0f9f-4f22-83a6-66c9e0942930-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.433190 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8dpc\" (UniqueName: \"kubernetes.io/projected/27d30ca5-0f9f-4f22-83a6-66c9e0942930-kube-api-access-g8dpc\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.471710 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-7d558687d7-q6tr2" podUID="5aaee075-cdd4-4532-a651-539dfc8b6ac9" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.161:9696/\": read tcp 10.217.0.2:38730->10.217.0.161:9696: read: connection reset by peer" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.534378 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d30ca5-0f9f-4f22-83a6-66c9e0942930-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.534440 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d7e90bc-8bc2-4c06-9131-405fa21670df-combined-ca-bundle\") pod \"neutron-76bd9cc6f7-x2p6w\" (UID: \"7d7e90bc-8bc2-4c06-9131-405fa21670df\") " pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.534478 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8dpc\" (UniqueName: \"kubernetes.io/projected/27d30ca5-0f9f-4f22-83a6-66c9e0942930-kube-api-access-g8dpc\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.534508 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7d7e90bc-8bc2-4c06-9131-405fa21670df-config\") pod \"neutron-76bd9cc6f7-x2p6w\" (UID: \"7d7e90bc-8bc2-4c06-9131-405fa21670df\") " pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.534540 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7d7e90bc-8bc2-4c06-9131-405fa21670df-httpd-config\") pod \"neutron-76bd9cc6f7-x2p6w\" (UID: \"7d7e90bc-8bc2-4c06-9131-405fa21670df\") " pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.534568 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jg9t\" (UniqueName: \"kubernetes.io/projected/7d7e90bc-8bc2-4c06-9131-405fa21670df-kube-api-access-5jg9t\") pod \"neutron-76bd9cc6f7-x2p6w\" (UID: \"7d7e90bc-8bc2-4c06-9131-405fa21670df\") " pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.534640 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27d30ca5-0f9f-4f22-83a6-66c9e0942930-logs\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.534674 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d7e90bc-8bc2-4c06-9131-405fa21670df-internal-tls-certs\") pod \"neutron-76bd9cc6f7-x2p6w\" (UID: \"7d7e90bc-8bc2-4c06-9131-405fa21670df\") " pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.534703 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27d30ca5-0f9f-4f22-83a6-66c9e0942930-config-data\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.535392 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27d30ca5-0f9f-4f22-83a6-66c9e0942930-logs\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.535577 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d7e90bc-8bc2-4c06-9131-405fa21670df-ovndb-tls-certs\") pod \"neutron-76bd9cc6f7-x2p6w\" (UID: \"7d7e90bc-8bc2-4c06-9131-405fa21670df\") " pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.535655 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d30ca5-0f9f-4f22-83a6-66c9e0942930-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.535697 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27d30ca5-0f9f-4f22-83a6-66c9e0942930-config-data-custom\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.535727 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27d30ca5-0f9f-4f22-83a6-66c9e0942930-scripts\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.535769 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d30ca5-0f9f-4f22-83a6-66c9e0942930-public-tls-certs\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.535805 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d7e90bc-8bc2-4c06-9131-405fa21670df-public-tls-certs\") pod \"neutron-76bd9cc6f7-x2p6w\" (UID: \"7d7e90bc-8bc2-4c06-9131-405fa21670df\") " pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.535829 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/27d30ca5-0f9f-4f22-83a6-66c9e0942930-etc-machine-id\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.535910 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/27d30ca5-0f9f-4f22-83a6-66c9e0942930-etc-machine-id\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.540646 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d30ca5-0f9f-4f22-83a6-66c9e0942930-public-tls-certs\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.540754 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27d30ca5-0f9f-4f22-83a6-66c9e0942930-config-data-custom\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.541342 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27d30ca5-0f9f-4f22-83a6-66c9e0942930-config-data\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.544166 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27d30ca5-0f9f-4f22-83a6-66c9e0942930-scripts\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.544519 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27d30ca5-0f9f-4f22-83a6-66c9e0942930-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.549465 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27d30ca5-0f9f-4f22-83a6-66c9e0942930-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.563126 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8dpc\" (UniqueName: \"kubernetes.io/projected/27d30ca5-0f9f-4f22-83a6-66c9e0942930-kube-api-access-g8dpc\") pod \"cinder-api-0\" (UID: \"27d30ca5-0f9f-4f22-83a6-66c9e0942930\") " pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.602921 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.642048 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/7d7e90bc-8bc2-4c06-9131-405fa21670df-config\") pod \"neutron-76bd9cc6f7-x2p6w\" (UID: \"7d7e90bc-8bc2-4c06-9131-405fa21670df\") " pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.642089 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7d7e90bc-8bc2-4c06-9131-405fa21670df-httpd-config\") pod \"neutron-76bd9cc6f7-x2p6w\" (UID: \"7d7e90bc-8bc2-4c06-9131-405fa21670df\") " pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.642112 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jg9t\" (UniqueName: \"kubernetes.io/projected/7d7e90bc-8bc2-4c06-9131-405fa21670df-kube-api-access-5jg9t\") pod \"neutron-76bd9cc6f7-x2p6w\" (UID: \"7d7e90bc-8bc2-4c06-9131-405fa21670df\") " pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.642168 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d7e90bc-8bc2-4c06-9131-405fa21670df-internal-tls-certs\") pod \"neutron-76bd9cc6f7-x2p6w\" (UID: \"7d7e90bc-8bc2-4c06-9131-405fa21670df\") " pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.642205 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d7e90bc-8bc2-4c06-9131-405fa21670df-ovndb-tls-certs\") pod \"neutron-76bd9cc6f7-x2p6w\" (UID: \"7d7e90bc-8bc2-4c06-9131-405fa21670df\") " pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.642260 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d7e90bc-8bc2-4c06-9131-405fa21670df-public-tls-certs\") pod \"neutron-76bd9cc6f7-x2p6w\" (UID: \"7d7e90bc-8bc2-4c06-9131-405fa21670df\") " pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.642299 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d7e90bc-8bc2-4c06-9131-405fa21670df-combined-ca-bundle\") pod \"neutron-76bd9cc6f7-x2p6w\" (UID: \"7d7e90bc-8bc2-4c06-9131-405fa21670df\") " pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.647537 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/7d7e90bc-8bc2-4c06-9131-405fa21670df-config\") pod \"neutron-76bd9cc6f7-x2p6w\" (UID: \"7d7e90bc-8bc2-4c06-9131-405fa21670df\") " pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.648281 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/7d7e90bc-8bc2-4c06-9131-405fa21670df-httpd-config\") pod \"neutron-76bd9cc6f7-x2p6w\" (UID: \"7d7e90bc-8bc2-4c06-9131-405fa21670df\") " pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.650365 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d7e90bc-8bc2-4c06-9131-405fa21670df-ovndb-tls-certs\") pod \"neutron-76bd9cc6f7-x2p6w\" (UID: \"7d7e90bc-8bc2-4c06-9131-405fa21670df\") " pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.650968 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d7e90bc-8bc2-4c06-9131-405fa21670df-public-tls-certs\") pod \"neutron-76bd9cc6f7-x2p6w\" (UID: \"7d7e90bc-8bc2-4c06-9131-405fa21670df\") " pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.651116 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7d7e90bc-8bc2-4c06-9131-405fa21670df-internal-tls-certs\") pod \"neutron-76bd9cc6f7-x2p6w\" (UID: \"7d7e90bc-8bc2-4c06-9131-405fa21670df\") " pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.651574 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d7e90bc-8bc2-4c06-9131-405fa21670df-combined-ca-bundle\") pod \"neutron-76bd9cc6f7-x2p6w\" (UID: \"7d7e90bc-8bc2-4c06-9131-405fa21670df\") " pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.669957 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jg9t\" (UniqueName: \"kubernetes.io/projected/7d7e90bc-8bc2-4c06-9131-405fa21670df-kube-api-access-5jg9t\") pod \"neutron-76bd9cc6f7-x2p6w\" (UID: \"7d7e90bc-8bc2-4c06-9131-405fa21670df\") " pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.690902 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.789868 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b8d767f8c-4ql8v" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.847506 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-config-data\") pod \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\" (UID: \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\") " Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.847572 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-logs\") pod \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\" (UID: \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\") " Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.847824 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqbwg\" (UniqueName: \"kubernetes.io/projected/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-kube-api-access-gqbwg\") pod \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\" (UID: \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\") " Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.847869 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-scripts\") pod \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\" (UID: \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\") " Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.847893 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-horizon-secret-key\") pod \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\" (UID: \"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6\") " Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.852068 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-logs" (OuterVolumeSpecName: "logs") pod "8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6" (UID: "8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.856755 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-kube-api-access-gqbwg" (OuterVolumeSpecName: "kube-api-access-gqbwg") pod "8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6" (UID: "8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6"). InnerVolumeSpecName "kube-api-access-gqbwg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.858714 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6" (UID: "8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.885435 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-scripts" (OuterVolumeSpecName: "scripts") pod "8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6" (UID: "8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.887065 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-config-data" (OuterVolumeSpecName: "config-data") pod "8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6" (UID: "8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.949523 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-logs\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.949550 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqbwg\" (UniqueName: \"kubernetes.io/projected/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-kube-api-access-gqbwg\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.949566 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.949575 4813 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:51 crc kubenswrapper[4813]: I0317 09:31:51.949584 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.179938 4813 generic.go:334] "Generic (PLEG): container finished" podID="8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6" containerID="c2963271c70c0c8b52d4a2447d8116c4b8f85c2c4475cacbe8918e346d55d5a4" exitCode=137 Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.179986 4813 generic.go:334] "Generic (PLEG): container finished" podID="8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6" containerID="d02f3c287e14f9b1f7666d914bf5a8c6fbd8dafffadba75dc06c5be99f1b9a19" exitCode=137 Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.179989 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8d767f8c-4ql8v" event={"ID":"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6","Type":"ContainerDied","Data":"c2963271c70c0c8b52d4a2447d8116c4b8f85c2c4475cacbe8918e346d55d5a4"} Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.180035 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8d767f8c-4ql8v" event={"ID":"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6","Type":"ContainerDied","Data":"d02f3c287e14f9b1f7666d914bf5a8c6fbd8dafffadba75dc06c5be99f1b9a19"} Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.180039 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7b8d767f8c-4ql8v" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.180057 4813 scope.go:117] "RemoveContainer" containerID="c2963271c70c0c8b52d4a2447d8116c4b8f85c2c4475cacbe8918e346d55d5a4" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.180045 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7b8d767f8c-4ql8v" event={"ID":"8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6","Type":"ContainerDied","Data":"7f903dc84603d24335df664545e8281ff4a8058a5e39da023b0f8738a752c373"} Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.183352 4813 generic.go:334] "Generic (PLEG): container finished" podID="5aaee075-cdd4-4532-a651-539dfc8b6ac9" containerID="f4c6cd7b65fba4bb7d14a208ed65138752b3610ba036e75b33bc344a062202be" exitCode=0 Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.183457 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d558687d7-q6tr2" event={"ID":"5aaee075-cdd4-4532-a651-539dfc8b6ac9","Type":"ContainerDied","Data":"f4c6cd7b65fba4bb7d14a208ed65138752b3610ba036e75b33bc344a062202be"} Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.232198 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.242498 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7b8d767f8c-4ql8v"] Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.249614 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7b8d767f8c-4ql8v"] Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.353368 4813 scope.go:117] "RemoveContainer" containerID="d02f3c287e14f9b1f7666d914bf5a8c6fbd8dafffadba75dc06c5be99f1b9a19" Mar 17 09:31:52 crc kubenswrapper[4813]: W0317 09:31:52.361180 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27d30ca5_0f9f_4f22_83a6_66c9e0942930.slice/crio-d69e9facddff00f0d0b8e722c61eb368800b5a9ba812f9775b1a8b03276f9125 WatchSource:0}: Error finding container d69e9facddff00f0d0b8e722c61eb368800b5a9ba812f9775b1a8b03276f9125: Status 404 returned error can't find the container with id d69e9facddff00f0d0b8e722c61eb368800b5a9ba812f9775b1a8b03276f9125 Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.378908 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-76bd9cc6f7-x2p6w"] Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.408377 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.506229 4813 scope.go:117] "RemoveContainer" containerID="c2963271c70c0c8b52d4a2447d8116c4b8f85c2c4475cacbe8918e346d55d5a4" Mar 17 09:31:52 crc kubenswrapper[4813]: E0317 09:31:52.506778 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2963271c70c0c8b52d4a2447d8116c4b8f85c2c4475cacbe8918e346d55d5a4\": container with ID starting with c2963271c70c0c8b52d4a2447d8116c4b8f85c2c4475cacbe8918e346d55d5a4 not found: ID does not exist" containerID="c2963271c70c0c8b52d4a2447d8116c4b8f85c2c4475cacbe8918e346d55d5a4" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.506816 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2963271c70c0c8b52d4a2447d8116c4b8f85c2c4475cacbe8918e346d55d5a4"} err="failed to get container status \"c2963271c70c0c8b52d4a2447d8116c4b8f85c2c4475cacbe8918e346d55d5a4\": rpc error: code = NotFound desc = could not find container \"c2963271c70c0c8b52d4a2447d8116c4b8f85c2c4475cacbe8918e346d55d5a4\": container with ID starting with c2963271c70c0c8b52d4a2447d8116c4b8f85c2c4475cacbe8918e346d55d5a4 not found: ID does not exist" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.506841 4813 scope.go:117] "RemoveContainer" containerID="d02f3c287e14f9b1f7666d914bf5a8c6fbd8dafffadba75dc06c5be99f1b9a19" Mar 17 09:31:52 crc kubenswrapper[4813]: E0317 09:31:52.507752 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d02f3c287e14f9b1f7666d914bf5a8c6fbd8dafffadba75dc06c5be99f1b9a19\": container with ID starting with d02f3c287e14f9b1f7666d914bf5a8c6fbd8dafffadba75dc06c5be99f1b9a19 not found: ID does not exist" containerID="d02f3c287e14f9b1f7666d914bf5a8c6fbd8dafffadba75dc06c5be99f1b9a19" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.507940 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d02f3c287e14f9b1f7666d914bf5a8c6fbd8dafffadba75dc06c5be99f1b9a19"} err="failed to get container status \"d02f3c287e14f9b1f7666d914bf5a8c6fbd8dafffadba75dc06c5be99f1b9a19\": rpc error: code = NotFound desc = could not find container \"d02f3c287e14f9b1f7666d914bf5a8c6fbd8dafffadba75dc06c5be99f1b9a19\": container with ID starting with d02f3c287e14f9b1f7666d914bf5a8c6fbd8dafffadba75dc06c5be99f1b9a19 not found: ID does not exist" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.508043 4813 scope.go:117] "RemoveContainer" containerID="c2963271c70c0c8b52d4a2447d8116c4b8f85c2c4475cacbe8918e346d55d5a4" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.508838 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2963271c70c0c8b52d4a2447d8116c4b8f85c2c4475cacbe8918e346d55d5a4"} err="failed to get container status \"c2963271c70c0c8b52d4a2447d8116c4b8f85c2c4475cacbe8918e346d55d5a4\": rpc error: code = NotFound desc = could not find container \"c2963271c70c0c8b52d4a2447d8116c4b8f85c2c4475cacbe8918e346d55d5a4\": container with ID starting with c2963271c70c0c8b52d4a2447d8116c4b8f85c2c4475cacbe8918e346d55d5a4 not found: ID does not exist" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.508872 4813 scope.go:117] "RemoveContainer" containerID="d02f3c287e14f9b1f7666d914bf5a8c6fbd8dafffadba75dc06c5be99f1b9a19" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.509099 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d02f3c287e14f9b1f7666d914bf5a8c6fbd8dafffadba75dc06c5be99f1b9a19"} err="failed to get container status \"d02f3c287e14f9b1f7666d914bf5a8c6fbd8dafffadba75dc06c5be99f1b9a19\": rpc error: code = NotFound desc = could not find container \"d02f3c287e14f9b1f7666d914bf5a8c6fbd8dafffadba75dc06c5be99f1b9a19\": container with ID starting with d02f3c287e14f9b1f7666d914bf5a8c6fbd8dafffadba75dc06c5be99f1b9a19 not found: ID does not exist" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.595734 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.696811 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-79bdd687db-2rnkr"] Mar 17 09:31:52 crc kubenswrapper[4813]: E0317 09:31:52.697219 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6" containerName="horizon" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.697235 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6" containerName="horizon" Mar 17 09:31:52 crc kubenswrapper[4813]: E0317 09:31:52.697275 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6" containerName="horizon-log" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.697281 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6" containerName="horizon-log" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.697438 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6" containerName="horizon" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.697457 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6" containerName="horizon-log" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.698386 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.700808 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.702508 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.709479 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-79bdd687db-2rnkr"] Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.774738 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6" path="/var/lib/kubelet/pods/8d6552e2-4f08-4b1e-9b83-c14d1b6e68f6/volumes" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.780064 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae0c0042-8dd1-44dc-a09f-5fefa2106098" path="/var/lib/kubelet/pods/ae0c0042-8dd1-44dc-a09f-5fefa2106098/volumes" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.868484 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80441abc-02b7-4a55-b4b9-2c3c7adf8ed2-combined-ca-bundle\") pod \"barbican-api-79bdd687db-2rnkr\" (UID: \"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2\") " pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.868662 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80441abc-02b7-4a55-b4b9-2c3c7adf8ed2-config-data\") pod \"barbican-api-79bdd687db-2rnkr\" (UID: \"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2\") " pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.869197 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80441abc-02b7-4a55-b4b9-2c3c7adf8ed2-logs\") pod \"barbican-api-79bdd687db-2rnkr\" (UID: \"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2\") " pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.869429 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/80441abc-02b7-4a55-b4b9-2c3c7adf8ed2-internal-tls-certs\") pod \"barbican-api-79bdd687db-2rnkr\" (UID: \"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2\") " pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.869547 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx8t2\" (UniqueName: \"kubernetes.io/projected/80441abc-02b7-4a55-b4b9-2c3c7adf8ed2-kube-api-access-hx8t2\") pod \"barbican-api-79bdd687db-2rnkr\" (UID: \"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2\") " pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.869657 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/80441abc-02b7-4a55-b4b9-2c3c7adf8ed2-public-tls-certs\") pod \"barbican-api-79bdd687db-2rnkr\" (UID: \"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2\") " pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.869779 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80441abc-02b7-4a55-b4b9-2c3c7adf8ed2-config-data-custom\") pod \"barbican-api-79bdd687db-2rnkr\" (UID: \"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2\") " pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.971693 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80441abc-02b7-4a55-b4b9-2c3c7adf8ed2-config-data\") pod \"barbican-api-79bdd687db-2rnkr\" (UID: \"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2\") " pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.971737 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80441abc-02b7-4a55-b4b9-2c3c7adf8ed2-logs\") pod \"barbican-api-79bdd687db-2rnkr\" (UID: \"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2\") " pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.971802 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/80441abc-02b7-4a55-b4b9-2c3c7adf8ed2-internal-tls-certs\") pod \"barbican-api-79bdd687db-2rnkr\" (UID: \"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2\") " pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.971837 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx8t2\" (UniqueName: \"kubernetes.io/projected/80441abc-02b7-4a55-b4b9-2c3c7adf8ed2-kube-api-access-hx8t2\") pod \"barbican-api-79bdd687db-2rnkr\" (UID: \"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2\") " pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.971873 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/80441abc-02b7-4a55-b4b9-2c3c7adf8ed2-public-tls-certs\") pod \"barbican-api-79bdd687db-2rnkr\" (UID: \"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2\") " pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.971908 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80441abc-02b7-4a55-b4b9-2c3c7adf8ed2-config-data-custom\") pod \"barbican-api-79bdd687db-2rnkr\" (UID: \"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2\") " pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.971949 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80441abc-02b7-4a55-b4b9-2c3c7adf8ed2-combined-ca-bundle\") pod \"barbican-api-79bdd687db-2rnkr\" (UID: \"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2\") " pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.972907 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/80441abc-02b7-4a55-b4b9-2c3c7adf8ed2-logs\") pod \"barbican-api-79bdd687db-2rnkr\" (UID: \"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2\") " pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.981125 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/80441abc-02b7-4a55-b4b9-2c3c7adf8ed2-public-tls-certs\") pod \"barbican-api-79bdd687db-2rnkr\" (UID: \"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2\") " pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.981355 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/80441abc-02b7-4a55-b4b9-2c3c7adf8ed2-config-data-custom\") pod \"barbican-api-79bdd687db-2rnkr\" (UID: \"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2\") " pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.982130 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/80441abc-02b7-4a55-b4b9-2c3c7adf8ed2-combined-ca-bundle\") pod \"barbican-api-79bdd687db-2rnkr\" (UID: \"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2\") " pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.992915 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/80441abc-02b7-4a55-b4b9-2c3c7adf8ed2-internal-tls-certs\") pod \"barbican-api-79bdd687db-2rnkr\" (UID: \"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2\") " pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.995343 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/80441abc-02b7-4a55-b4b9-2c3c7adf8ed2-config-data\") pod \"barbican-api-79bdd687db-2rnkr\" (UID: \"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2\") " pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:52 crc kubenswrapper[4813]: I0317 09:31:52.998425 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx8t2\" (UniqueName: \"kubernetes.io/projected/80441abc-02b7-4a55-b4b9-2c3c7adf8ed2-kube-api-access-hx8t2\") pod \"barbican-api-79bdd687db-2rnkr\" (UID: \"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2\") " pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:53 crc kubenswrapper[4813]: I0317 09:31:53.039069 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:53 crc kubenswrapper[4813]: I0317 09:31:53.219522 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76bd9cc6f7-x2p6w" event={"ID":"7d7e90bc-8bc2-4c06-9131-405fa21670df","Type":"ContainerStarted","Data":"c99eadb01a65207ce792985a29b91985c2979c9b678e850d60e2354ea8d049ca"} Mar 17 09:31:53 crc kubenswrapper[4813]: I0317 09:31:53.219842 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:31:53 crc kubenswrapper[4813]: I0317 09:31:53.219858 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76bd9cc6f7-x2p6w" event={"ID":"7d7e90bc-8bc2-4c06-9131-405fa21670df","Type":"ContainerStarted","Data":"471c8f2bd6e57afe579e39e603583a6f453c6821428629fd46f111a3023ab4c5"} Mar 17 09:31:53 crc kubenswrapper[4813]: I0317 09:31:53.219870 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-76bd9cc6f7-x2p6w" event={"ID":"7d7e90bc-8bc2-4c06-9131-405fa21670df","Type":"ContainerStarted","Data":"bfe10b3e7c2ea9485e55bc3a06cd639e67f077f54763d1a714a679b6f3611ba3"} Mar 17 09:31:53 crc kubenswrapper[4813]: I0317 09:31:53.247144 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"27d30ca5-0f9f-4f22-83a6-66c9e0942930","Type":"ContainerStarted","Data":"d69e9facddff00f0d0b8e722c61eb368800b5a9ba812f9775b1a8b03276f9125"} Mar 17 09:31:53 crc kubenswrapper[4813]: I0317 09:31:53.252202 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-76bd9cc6f7-x2p6w" podStartSLOduration=2.25218435 podStartE2EDuration="2.25218435s" podCreationTimestamp="2026-03-17 09:31:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:31:53.241024969 +0000 UTC m=+1335.341828458" watchObservedRunningTime="2026-03-17 09:31:53.25218435 +0000 UTC m=+1335.352987849" Mar 17 09:31:53 crc kubenswrapper[4813]: I0317 09:31:53.368908 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/neutron-7d558687d7-q6tr2" podUID="5aaee075-cdd4-4532-a651-539dfc8b6ac9" containerName="neutron-httpd" probeResult="failure" output="Get \"https://10.217.0.161:9696/\": dial tcp 10.217.0.161:9696: connect: connection refused" Mar 17 09:31:53 crc kubenswrapper[4813]: I0317 09:31:53.687008 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-79bdd687db-2rnkr"] Mar 17 09:31:53 crc kubenswrapper[4813]: W0317 09:31:53.690323 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod80441abc_02b7_4a55_b4b9_2c3c7adf8ed2.slice/crio-9e68361565241767e117264849685f5429ed692a675bba765466f72fe8827b79 WatchSource:0}: Error finding container 9e68361565241767e117264849685f5429ed692a675bba765466f72fe8827b79: Status 404 returned error can't find the container with id 9e68361565241767e117264849685f5429ed692a675bba765466f72fe8827b79 Mar 17 09:31:54 crc kubenswrapper[4813]: I0317 09:31:54.257735 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"27d30ca5-0f9f-4f22-83a6-66c9e0942930","Type":"ContainerStarted","Data":"3b9f81d316cf3b67b6a8ccee7ed3a6f88d7e6084ae088d6d4dd894ba5d99482b"} Mar 17 09:31:54 crc kubenswrapper[4813]: I0317 09:31:54.260391 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79bdd687db-2rnkr" event={"ID":"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2","Type":"ContainerStarted","Data":"e0f0626fb114f13d2342b347144a2e9d74f204e2df7fdc27c0d544ec3ebebf72"} Mar 17 09:31:54 crc kubenswrapper[4813]: I0317 09:31:54.260430 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79bdd687db-2rnkr" event={"ID":"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2","Type":"ContainerStarted","Data":"9e68361565241767e117264849685f5429ed692a675bba765466f72fe8827b79"} Mar 17 09:31:54 crc kubenswrapper[4813]: I0317 09:31:54.658256 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:31:54 crc kubenswrapper[4813]: I0317 09:31:54.828463 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5dc999f6f8-t94dl" Mar 17 09:31:54 crc kubenswrapper[4813]: I0317 09:31:54.888085 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-85fc85fdfb-n469t"] Mar 17 09:31:55 crc kubenswrapper[4813]: I0317 09:31:55.269846 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-79bdd687db-2rnkr" event={"ID":"80441abc-02b7-4a55-b4b9-2c3c7adf8ed2","Type":"ContainerStarted","Data":"a4a314d9ebea6f474b48782aa9b6a3d80d2918c610a8d00361dff7cfa48cc0bd"} Mar 17 09:31:55 crc kubenswrapper[4813]: I0317 09:31:55.272441 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:55 crc kubenswrapper[4813]: I0317 09:31:55.272473 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:31:55 crc kubenswrapper[4813]: I0317 09:31:55.276319 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"27d30ca5-0f9f-4f22-83a6-66c9e0942930","Type":"ContainerStarted","Data":"b97daac5ad1d9d4bcb4db88a29b2c3da7e26257a598e24bae3918e95acb2771c"} Mar 17 09:31:55 crc kubenswrapper[4813]: I0317 09:31:55.276491 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-85fc85fdfb-n469t" podUID="0a52b24b-cc19-4720-a05b-961015362769" containerName="horizon-log" containerID="cri-o://f4b4b9a07ddaca0201104a913247a2025ba56488f789d4d1b11072bb55859ee4" gracePeriod=30 Mar 17 09:31:55 crc kubenswrapper[4813]: I0317 09:31:55.276756 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-85fc85fdfb-n469t" podUID="0a52b24b-cc19-4720-a05b-961015362769" containerName="horizon" containerID="cri-o://77e6158b584698861562f3c366b15b1b09acecaf7c7eb2ee5edd3c9ddec18130" gracePeriod=30 Mar 17 09:31:55 crc kubenswrapper[4813]: I0317 09:31:55.326639 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-79bdd687db-2rnkr" podStartSLOduration=3.326622372 podStartE2EDuration="3.326622372s" podCreationTimestamp="2026-03-17 09:31:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:31:55.298518148 +0000 UTC m=+1337.399321667" watchObservedRunningTime="2026-03-17 09:31:55.326622372 +0000 UTC m=+1337.427425871" Mar 17 09:31:55 crc kubenswrapper[4813]: I0317 09:31:55.331386 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.331371362 podStartE2EDuration="4.331371362s" podCreationTimestamp="2026-03-17 09:31:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:31:55.324505856 +0000 UTC m=+1337.425309345" watchObservedRunningTime="2026-03-17 09:31:55.331371362 +0000 UTC m=+1337.432174861" Mar 17 09:31:56 crc kubenswrapper[4813]: I0317 09:31:56.084250 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 17 09:31:56 crc kubenswrapper[4813]: I0317 09:31:56.135516 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 17 09:31:56 crc kubenswrapper[4813]: I0317 09:31:56.284102 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="3a5534d6-3502-4a49-8991-cf7fe05d2e8d" containerName="cinder-scheduler" containerID="cri-o://f1396ad8c478422139fb9b0205ca8715ff597cfab4f6cd9e69e2aa70bb114796" gracePeriod=30 Mar 17 09:31:56 crc kubenswrapper[4813]: I0317 09:31:56.286919 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="3a5534d6-3502-4a49-8991-cf7fe05d2e8d" containerName="probe" containerID="cri-o://50b311fe853f38caed37e1ab6162fade6400373689947e62efdaea2a928d8569" gracePeriod=30 Mar 17 09:31:56 crc kubenswrapper[4813]: I0317 09:31:56.288564 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Mar 17 09:31:56 crc kubenswrapper[4813]: I0317 09:31:56.324262 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:31:56 crc kubenswrapper[4813]: I0317 09:31:56.424959 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-mfck7"] Mar 17 09:31:56 crc kubenswrapper[4813]: I0317 09:31:56.425202 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" podUID="de2cccf5-7797-4baf-bf2f-d6405959799f" containerName="dnsmasq-dns" containerID="cri-o://1b99d283d11cd2d81eba803093711bc717ff99fd015ca120f035046e76ba25dd" gracePeriod=10 Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.305538 4813 generic.go:334] "Generic (PLEG): container finished" podID="5aaee075-cdd4-4532-a651-539dfc8b6ac9" containerID="8319e5f3fa73599000d2c08ec05ceacf5dcc9f68e9da36049d345fa482790c15" exitCode=0 Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.305636 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d558687d7-q6tr2" event={"ID":"5aaee075-cdd4-4532-a651-539dfc8b6ac9","Type":"ContainerDied","Data":"8319e5f3fa73599000d2c08ec05ceacf5dcc9f68e9da36049d345fa482790c15"} Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.314226 4813 generic.go:334] "Generic (PLEG): container finished" podID="de2cccf5-7797-4baf-bf2f-d6405959799f" containerID="1b99d283d11cd2d81eba803093711bc717ff99fd015ca120f035046e76ba25dd" exitCode=0 Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.314305 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" event={"ID":"de2cccf5-7797-4baf-bf2f-d6405959799f","Type":"ContainerDied","Data":"1b99d283d11cd2d81eba803093711bc717ff99fd015ca120f035046e76ba25dd"} Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.320930 4813 generic.go:334] "Generic (PLEG): container finished" podID="3a5534d6-3502-4a49-8991-cf7fe05d2e8d" containerID="50b311fe853f38caed37e1ab6162fade6400373689947e62efdaea2a928d8569" exitCode=0 Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.321015 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3a5534d6-3502-4a49-8991-cf7fe05d2e8d","Type":"ContainerDied","Data":"50b311fe853f38caed37e1ab6162fade6400373689947e62efdaea2a928d8569"} Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.429295 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.535712 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.663646 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.788403 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-557566c676-bbf8q"] Mar 17 09:31:57 crc kubenswrapper[4813]: E0317 09:31:57.788975 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de2cccf5-7797-4baf-bf2f-d6405959799f" containerName="init" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.788986 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="de2cccf5-7797-4baf-bf2f-d6405959799f" containerName="init" Mar 17 09:31:57 crc kubenswrapper[4813]: E0317 09:31:57.789010 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de2cccf5-7797-4baf-bf2f-d6405959799f" containerName="dnsmasq-dns" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.789016 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="de2cccf5-7797-4baf-bf2f-d6405959799f" containerName="dnsmasq-dns" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.789171 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="de2cccf5-7797-4baf-bf2f-d6405959799f" containerName="dnsmasq-dns" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.790033 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.809323 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-ovsdbserver-nb\") pod \"de2cccf5-7797-4baf-bf2f-d6405959799f\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.809483 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r62cj\" (UniqueName: \"kubernetes.io/projected/de2cccf5-7797-4baf-bf2f-d6405959799f-kube-api-access-r62cj\") pod \"de2cccf5-7797-4baf-bf2f-d6405959799f\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.809510 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-ovsdbserver-sb\") pod \"de2cccf5-7797-4baf-bf2f-d6405959799f\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.809577 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-config\") pod \"de2cccf5-7797-4baf-bf2f-d6405959799f\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.809614 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-dns-swift-storage-0\") pod \"de2cccf5-7797-4baf-bf2f-d6405959799f\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.809682 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-dns-svc\") pod \"de2cccf5-7797-4baf-bf2f-d6405959799f\" (UID: \"de2cccf5-7797-4baf-bf2f-d6405959799f\") " Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.846059 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de2cccf5-7797-4baf-bf2f-d6405959799f-kube-api-access-r62cj" (OuterVolumeSpecName: "kube-api-access-r62cj") pod "de2cccf5-7797-4baf-bf2f-d6405959799f" (UID: "de2cccf5-7797-4baf-bf2f-d6405959799f"). InnerVolumeSpecName "kube-api-access-r62cj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.862250 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-557566c676-bbf8q"] Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.890890 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "de2cccf5-7797-4baf-bf2f-d6405959799f" (UID: "de2cccf5-7797-4baf-bf2f-d6405959799f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.907274 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "de2cccf5-7797-4baf-bf2f-d6405959799f" (UID: "de2cccf5-7797-4baf-bf2f-d6405959799f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.911868 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w6sxt\" (UniqueName: \"kubernetes.io/projected/46eacf89-88ff-4be7-a8c1-e90784324da2-kube-api-access-w6sxt\") pod \"placement-557566c676-bbf8q\" (UID: \"46eacf89-88ff-4be7-a8c1-e90784324da2\") " pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.911913 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46eacf89-88ff-4be7-a8c1-e90784324da2-combined-ca-bundle\") pod \"placement-557566c676-bbf8q\" (UID: \"46eacf89-88ff-4be7-a8c1-e90784324da2\") " pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.912075 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/46eacf89-88ff-4be7-a8c1-e90784324da2-public-tls-certs\") pod \"placement-557566c676-bbf8q\" (UID: \"46eacf89-88ff-4be7-a8c1-e90784324da2\") " pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.912201 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46eacf89-88ff-4be7-a8c1-e90784324da2-config-data\") pod \"placement-557566c676-bbf8q\" (UID: \"46eacf89-88ff-4be7-a8c1-e90784324da2\") " pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.912244 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46eacf89-88ff-4be7-a8c1-e90784324da2-internal-tls-certs\") pod \"placement-557566c676-bbf8q\" (UID: \"46eacf89-88ff-4be7-a8c1-e90784324da2\") " pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.912297 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46eacf89-88ff-4be7-a8c1-e90784324da2-logs\") pod \"placement-557566c676-bbf8q\" (UID: \"46eacf89-88ff-4be7-a8c1-e90784324da2\") " pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.912322 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46eacf89-88ff-4be7-a8c1-e90784324da2-scripts\") pod \"placement-557566c676-bbf8q\" (UID: \"46eacf89-88ff-4be7-a8c1-e90784324da2\") " pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.912403 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.912420 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r62cj\" (UniqueName: \"kubernetes.io/projected/de2cccf5-7797-4baf-bf2f-d6405959799f-kube-api-access-r62cj\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.912430 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.923327 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-config" (OuterVolumeSpecName: "config") pod "de2cccf5-7797-4baf-bf2f-d6405959799f" (UID: "de2cccf5-7797-4baf-bf2f-d6405959799f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.955821 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "de2cccf5-7797-4baf-bf2f-d6405959799f" (UID: "de2cccf5-7797-4baf-bf2f-d6405959799f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.993056 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:57 crc kubenswrapper[4813]: I0317 09:31:57.995074 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "de2cccf5-7797-4baf-bf2f-d6405959799f" (UID: "de2cccf5-7797-4baf-bf2f-d6405959799f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.016561 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w6sxt\" (UniqueName: \"kubernetes.io/projected/46eacf89-88ff-4be7-a8c1-e90784324da2-kube-api-access-w6sxt\") pod \"placement-557566c676-bbf8q\" (UID: \"46eacf89-88ff-4be7-a8c1-e90784324da2\") " pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.016702 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46eacf89-88ff-4be7-a8c1-e90784324da2-combined-ca-bundle\") pod \"placement-557566c676-bbf8q\" (UID: \"46eacf89-88ff-4be7-a8c1-e90784324da2\") " pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.016765 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/46eacf89-88ff-4be7-a8c1-e90784324da2-public-tls-certs\") pod \"placement-557566c676-bbf8q\" (UID: \"46eacf89-88ff-4be7-a8c1-e90784324da2\") " pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.016807 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46eacf89-88ff-4be7-a8c1-e90784324da2-config-data\") pod \"placement-557566c676-bbf8q\" (UID: \"46eacf89-88ff-4be7-a8c1-e90784324da2\") " pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.016827 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46eacf89-88ff-4be7-a8c1-e90784324da2-internal-tls-certs\") pod \"placement-557566c676-bbf8q\" (UID: \"46eacf89-88ff-4be7-a8c1-e90784324da2\") " pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.016852 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46eacf89-88ff-4be7-a8c1-e90784324da2-logs\") pod \"placement-557566c676-bbf8q\" (UID: \"46eacf89-88ff-4be7-a8c1-e90784324da2\") " pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.016868 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46eacf89-88ff-4be7-a8c1-e90784324da2-scripts\") pod \"placement-557566c676-bbf8q\" (UID: \"46eacf89-88ff-4be7-a8c1-e90784324da2\") " pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.016923 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.016933 4813 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.016944 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/de2cccf5-7797-4baf-bf2f-d6405959799f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.019917 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/46eacf89-88ff-4be7-a8c1-e90784324da2-logs\") pod \"placement-557566c676-bbf8q\" (UID: \"46eacf89-88ff-4be7-a8c1-e90784324da2\") " pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.024557 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46eacf89-88ff-4be7-a8c1-e90784324da2-scripts\") pod \"placement-557566c676-bbf8q\" (UID: \"46eacf89-88ff-4be7-a8c1-e90784324da2\") " pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.024943 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.033489 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46eacf89-88ff-4be7-a8c1-e90784324da2-config-data\") pod \"placement-557566c676-bbf8q\" (UID: \"46eacf89-88ff-4be7-a8c1-e90784324da2\") " pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.034259 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/46eacf89-88ff-4be7-a8c1-e90784324da2-internal-tls-certs\") pod \"placement-557566c676-bbf8q\" (UID: \"46eacf89-88ff-4be7-a8c1-e90784324da2\") " pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.034698 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46eacf89-88ff-4be7-a8c1-e90784324da2-combined-ca-bundle\") pod \"placement-557566c676-bbf8q\" (UID: \"46eacf89-88ff-4be7-a8c1-e90784324da2\") " pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.037171 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/46eacf89-88ff-4be7-a8c1-e90784324da2-public-tls-certs\") pod \"placement-557566c676-bbf8q\" (UID: \"46eacf89-88ff-4be7-a8c1-e90784324da2\") " pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.042803 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w6sxt\" (UniqueName: \"kubernetes.io/projected/46eacf89-88ff-4be7-a8c1-e90784324da2-kube-api-access-w6sxt\") pod \"placement-557566c676-bbf8q\" (UID: \"46eacf89-88ff-4be7-a8c1-e90784324da2\") " pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.118146 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-config\") pod \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.118211 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-httpd-config\") pod \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.118264 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-public-tls-certs\") pod \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.118281 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-internal-tls-certs\") pod \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.118323 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-combined-ca-bundle\") pod \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.118386 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dtgff\" (UniqueName: \"kubernetes.io/projected/5aaee075-cdd4-4532-a651-539dfc8b6ac9-kube-api-access-dtgff\") pod \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.118413 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-ovndb-tls-certs\") pod \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\" (UID: \"5aaee075-cdd4-4532-a651-539dfc8b6ac9\") " Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.124787 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5aaee075-cdd4-4532-a651-539dfc8b6ac9-kube-api-access-dtgff" (OuterVolumeSpecName: "kube-api-access-dtgff") pod "5aaee075-cdd4-4532-a651-539dfc8b6ac9" (UID: "5aaee075-cdd4-4532-a651-539dfc8b6ac9"). InnerVolumeSpecName "kube-api-access-dtgff". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.124789 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "5aaee075-cdd4-4532-a651-539dfc8b6ac9" (UID: "5aaee075-cdd4-4532-a651-539dfc8b6ac9"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.174878 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "5aaee075-cdd4-4532-a651-539dfc8b6ac9" (UID: "5aaee075-cdd4-4532-a651-539dfc8b6ac9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.185101 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-config" (OuterVolumeSpecName: "config") pod "5aaee075-cdd4-4532-a651-539dfc8b6ac9" (UID: "5aaee075-cdd4-4532-a651-539dfc8b6ac9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.199092 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5aaee075-cdd4-4532-a651-539dfc8b6ac9" (UID: "5aaee075-cdd4-4532-a651-539dfc8b6ac9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.208859 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "5aaee075-cdd4-4532-a651-539dfc8b6ac9" (UID: "5aaee075-cdd4-4532-a651-539dfc8b6ac9"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.221176 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.221208 4813 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.221218 4813 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.221229 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.221237 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dtgff\" (UniqueName: \"kubernetes.io/projected/5aaee075-cdd4-4532-a651-539dfc8b6ac9-kube-api-access-dtgff\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.221248 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.230739 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "5aaee075-cdd4-4532-a651-539dfc8b6ac9" (UID: "5aaee075-cdd4-4532-a651-539dfc8b6ac9"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.286187 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.322589 4813 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/5aaee075-cdd4-4532-a651-539dfc8b6ac9-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.350432 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.404020 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" event={"ID":"de2cccf5-7797-4baf-bf2f-d6405959799f","Type":"ContainerDied","Data":"c37d2c93e6802dd7b7fbc5992f764c8658415869afaaf7df166092667ace066b"} Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.404104 4813 scope.go:117] "RemoveContainer" containerID="1b99d283d11cd2d81eba803093711bc717ff99fd015ca120f035046e76ba25dd" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.404050 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84b966f6c9-mfck7" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.468700 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-7d558687d7-q6tr2" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.471117 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-7d558687d7-q6tr2" event={"ID":"5aaee075-cdd4-4532-a651-539dfc8b6ac9","Type":"ContainerDied","Data":"cecfa89ed18ca4e06996dd45390bc4db8bdab3bd956c41a07ba7e58e30e9710d"} Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.478830 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-mfck7"] Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.479757 4813 scope.go:117] "RemoveContainer" containerID="b963f4bff251784bbbfa589d509487039ab20258d62b37f6cfe465f23047cca6" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.492194 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84b966f6c9-mfck7"] Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.520642 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-7d558687d7-q6tr2"] Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.527997 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-7d558687d7-q6tr2"] Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.598851 4813 scope.go:117] "RemoveContainer" containerID="f4c6cd7b65fba4bb7d14a208ed65138752b3610ba036e75b33bc344a062202be" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.699193 4813 scope.go:117] "RemoveContainer" containerID="8319e5f3fa73599000d2c08ec05ceacf5dcc9f68e9da36049d345fa482790c15" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.750313 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5aaee075-cdd4-4532-a651-539dfc8b6ac9" path="/var/lib/kubelet/pods/5aaee075-cdd4-4532-a651-539dfc8b6ac9/volumes" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.750884 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de2cccf5-7797-4baf-bf2f-d6405959799f" path="/var/lib/kubelet/pods/de2cccf5-7797-4baf-bf2f-d6405959799f/volumes" Mar 17 09:31:58 crc kubenswrapper[4813]: I0317 09:31:58.929232 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-557566c676-bbf8q"] Mar 17 09:31:58 crc kubenswrapper[4813]: W0317 09:31:58.949674 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46eacf89_88ff_4be7_a8c1_e90784324da2.slice/crio-2fa13f703a3c4f20d2d31420161f2652204abef1cdb542992ac5a67d9bec2080 WatchSource:0}: Error finding container 2fa13f703a3c4f20d2d31420161f2652204abef1cdb542992ac5a67d9bec2080: Status 404 returned error can't find the container with id 2fa13f703a3c4f20d2d31420161f2652204abef1cdb542992ac5a67d9bec2080 Mar 17 09:31:59 crc kubenswrapper[4813]: I0317 09:31:59.475080 4813 generic.go:334] "Generic (PLEG): container finished" podID="0a52b24b-cc19-4720-a05b-961015362769" containerID="77e6158b584698861562f3c366b15b1b09acecaf7c7eb2ee5edd3c9ddec18130" exitCode=0 Mar 17 09:31:59 crc kubenswrapper[4813]: I0317 09:31:59.475279 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85fc85fdfb-n469t" event={"ID":"0a52b24b-cc19-4720-a05b-961015362769","Type":"ContainerDied","Data":"77e6158b584698861562f3c366b15b1b09acecaf7c7eb2ee5edd3c9ddec18130"} Mar 17 09:31:59 crc kubenswrapper[4813]: I0317 09:31:59.477783 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-557566c676-bbf8q" event={"ID":"46eacf89-88ff-4be7-a8c1-e90784324da2","Type":"ContainerStarted","Data":"47e5c5e831b2ca16b785e95cbed483ec13e3f0c366c80d600b8c4e8a075a2947"} Mar 17 09:31:59 crc kubenswrapper[4813]: I0317 09:31:59.477813 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-557566c676-bbf8q" event={"ID":"46eacf89-88ff-4be7-a8c1-e90784324da2","Type":"ContainerStarted","Data":"8b306fa2d00d2715f75ff9c9e146a4429a1da4df2e2423fbc8ab39244dfd9c19"} Mar 17 09:31:59 crc kubenswrapper[4813]: I0317 09:31:59.477825 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-557566c676-bbf8q" event={"ID":"46eacf89-88ff-4be7-a8c1-e90784324da2","Type":"ContainerStarted","Data":"2fa13f703a3c4f20d2d31420161f2652204abef1cdb542992ac5a67d9bec2080"} Mar 17 09:31:59 crc kubenswrapper[4813]: I0317 09:31:59.478911 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:59 crc kubenswrapper[4813]: I0317 09:31:59.478961 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-557566c676-bbf8q" Mar 17 09:31:59 crc kubenswrapper[4813]: I0317 09:31:59.507134 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-557566c676-bbf8q" podStartSLOduration=2.507114292 podStartE2EDuration="2.507114292s" podCreationTimestamp="2026-03-17 09:31:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:31:59.494089053 +0000 UTC m=+1341.594892552" watchObservedRunningTime="2026-03-17 09:31:59.507114292 +0000 UTC m=+1341.607917791" Mar 17 09:31:59 crc kubenswrapper[4813]: I0317 09:31:59.858116 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:32:00 crc kubenswrapper[4813]: I0317 09:32:00.131614 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562332-p5phc"] Mar 17 09:32:00 crc kubenswrapper[4813]: E0317 09:32:00.131952 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aaee075-cdd4-4532-a651-539dfc8b6ac9" containerName="neutron-api" Mar 17 09:32:00 crc kubenswrapper[4813]: I0317 09:32:00.131963 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aaee075-cdd4-4532-a651-539dfc8b6ac9" containerName="neutron-api" Mar 17 09:32:00 crc kubenswrapper[4813]: E0317 09:32:00.132000 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5aaee075-cdd4-4532-a651-539dfc8b6ac9" containerName="neutron-httpd" Mar 17 09:32:00 crc kubenswrapper[4813]: I0317 09:32:00.132007 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5aaee075-cdd4-4532-a651-539dfc8b6ac9" containerName="neutron-httpd" Mar 17 09:32:00 crc kubenswrapper[4813]: I0317 09:32:00.132155 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5aaee075-cdd4-4532-a651-539dfc8b6ac9" containerName="neutron-api" Mar 17 09:32:00 crc kubenswrapper[4813]: I0317 09:32:00.132172 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5aaee075-cdd4-4532-a651-539dfc8b6ac9" containerName="neutron-httpd" Mar 17 09:32:00 crc kubenswrapper[4813]: I0317 09:32:00.132680 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562332-p5phc" Mar 17 09:32:00 crc kubenswrapper[4813]: I0317 09:32:00.140871 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:32:00 crc kubenswrapper[4813]: I0317 09:32:00.140946 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:32:00 crc kubenswrapper[4813]: I0317 09:32:00.142027 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:32:00 crc kubenswrapper[4813]: I0317 09:32:00.152555 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562332-p5phc"] Mar 17 09:32:00 crc kubenswrapper[4813]: I0317 09:32:00.257556 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5bnl\" (UniqueName: \"kubernetes.io/projected/4cf3e585-9c98-4e66-b577-59ddd1327ca2-kube-api-access-t5bnl\") pod \"auto-csr-approver-29562332-p5phc\" (UID: \"4cf3e585-9c98-4e66-b577-59ddd1327ca2\") " pod="openshift-infra/auto-csr-approver-29562332-p5phc" Mar 17 09:32:00 crc kubenswrapper[4813]: I0317 09:32:00.306262 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-85fc85fdfb-n469t" podUID="0a52b24b-cc19-4720-a05b-961015362769" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.156:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.156:8443: connect: connection refused" Mar 17 09:32:00 crc kubenswrapper[4813]: I0317 09:32:00.359708 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5bnl\" (UniqueName: \"kubernetes.io/projected/4cf3e585-9c98-4e66-b577-59ddd1327ca2-kube-api-access-t5bnl\") pod \"auto-csr-approver-29562332-p5phc\" (UID: \"4cf3e585-9c98-4e66-b577-59ddd1327ca2\") " pod="openshift-infra/auto-csr-approver-29562332-p5phc" Mar 17 09:32:00 crc kubenswrapper[4813]: I0317 09:32:00.402277 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5bnl\" (UniqueName: \"kubernetes.io/projected/4cf3e585-9c98-4e66-b577-59ddd1327ca2-kube-api-access-t5bnl\") pod \"auto-csr-approver-29562332-p5phc\" (UID: \"4cf3e585-9c98-4e66-b577-59ddd1327ca2\") " pod="openshift-infra/auto-csr-approver-29562332-p5phc" Mar 17 09:32:00 crc kubenswrapper[4813]: I0317 09:32:00.450318 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562332-p5phc" Mar 17 09:32:00 crc kubenswrapper[4813]: I0317 09:32:00.521971 4813 generic.go:334] "Generic (PLEG): container finished" podID="3a5534d6-3502-4a49-8991-cf7fe05d2e8d" containerID="f1396ad8c478422139fb9b0205ca8715ff597cfab4f6cd9e69e2aa70bb114796" exitCode=0 Mar 17 09:32:00 crc kubenswrapper[4813]: I0317 09:32:00.522847 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3a5534d6-3502-4a49-8991-cf7fe05d2e8d","Type":"ContainerDied","Data":"f1396ad8c478422139fb9b0205ca8715ff597cfab4f6cd9e69e2aa70bb114796"} Mar 17 09:32:00 crc kubenswrapper[4813]: I0317 09:32:00.978302 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.074519 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-etc-machine-id\") pod \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.074671 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-config-data\") pod \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.074698 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mj4k8\" (UniqueName: \"kubernetes.io/projected/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-kube-api-access-mj4k8\") pod \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.074754 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-scripts\") pod \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.074791 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-combined-ca-bundle\") pod \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.074807 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-config-data-custom\") pod \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\" (UID: \"3a5534d6-3502-4a49-8991-cf7fe05d2e8d\") " Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.075900 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "3a5534d6-3502-4a49-8991-cf7fe05d2e8d" (UID: "3a5534d6-3502-4a49-8991-cf7fe05d2e8d"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.079397 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562332-p5phc"] Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.082827 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-scripts" (OuterVolumeSpecName: "scripts") pod "3a5534d6-3502-4a49-8991-cf7fe05d2e8d" (UID: "3a5534d6-3502-4a49-8991-cf7fe05d2e8d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.082874 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "3a5534d6-3502-4a49-8991-cf7fe05d2e8d" (UID: "3a5534d6-3502-4a49-8991-cf7fe05d2e8d"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:01 crc kubenswrapper[4813]: W0317 09:32:01.083624 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4cf3e585_9c98_4e66_b577_59ddd1327ca2.slice/crio-6d819e4356134cbb82a3b591162f32a5f6122f65ebe6c008d14c32282e409b79 WatchSource:0}: Error finding container 6d819e4356134cbb82a3b591162f32a5f6122f65ebe6c008d14c32282e409b79: Status 404 returned error can't find the container with id 6d819e4356134cbb82a3b591162f32a5f6122f65ebe6c008d14c32282e409b79 Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.083552 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-kube-api-access-mj4k8" (OuterVolumeSpecName: "kube-api-access-mj4k8") pod "3a5534d6-3502-4a49-8991-cf7fe05d2e8d" (UID: "3a5534d6-3502-4a49-8991-cf7fe05d2e8d"). InnerVolumeSpecName "kube-api-access-mj4k8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.116094 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-7fb6796ddc-58j2l" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.142754 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3a5534d6-3502-4a49-8991-cf7fe05d2e8d" (UID: "3a5534d6-3502-4a49-8991-cf7fe05d2e8d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.177235 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mj4k8\" (UniqueName: \"kubernetes.io/projected/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-kube-api-access-mj4k8\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.177272 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.177285 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.177296 4813 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.177308 4813 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-etc-machine-id\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.226688 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-config-data" (OuterVolumeSpecName: "config-data") pod "3a5534d6-3502-4a49-8991-cf7fe05d2e8d" (UID: "3a5534d6-3502-4a49-8991-cf7fe05d2e8d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.278994 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3a5534d6-3502-4a49-8991-cf7fe05d2e8d-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.530856 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562332-p5phc" event={"ID":"4cf3e585-9c98-4e66-b577-59ddd1327ca2","Type":"ContainerStarted","Data":"6d819e4356134cbb82a3b591162f32a5f6122f65ebe6c008d14c32282e409b79"} Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.533398 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.533441 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"3a5534d6-3502-4a49-8991-cf7fe05d2e8d","Type":"ContainerDied","Data":"79a39c8324945965793609c773665e493204f54fe03d7d4c9b01ebb2a9f32165"} Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.533474 4813 scope.go:117] "RemoveContainer" containerID="50b311fe853f38caed37e1ab6162fade6400373689947e62efdaea2a928d8569" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.558117 4813 scope.go:117] "RemoveContainer" containerID="f1396ad8c478422139fb9b0205ca8715ff597cfab4f6cd9e69e2aa70bb114796" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.576758 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.583530 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.602235 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Mar 17 09:32:01 crc kubenswrapper[4813]: E0317 09:32:01.602797 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a5534d6-3502-4a49-8991-cf7fe05d2e8d" containerName="cinder-scheduler" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.602821 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a5534d6-3502-4a49-8991-cf7fe05d2e8d" containerName="cinder-scheduler" Mar 17 09:32:01 crc kubenswrapper[4813]: E0317 09:32:01.602881 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a5534d6-3502-4a49-8991-cf7fe05d2e8d" containerName="probe" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.602891 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a5534d6-3502-4a49-8991-cf7fe05d2e8d" containerName="probe" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.612093 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a5534d6-3502-4a49-8991-cf7fe05d2e8d" containerName="cinder-scheduler" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.612167 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a5534d6-3502-4a49-8991-cf7fe05d2e8d" containerName="probe" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.613476 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.616520 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.626155 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.685743 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2cad367-9c0f-4556-9e4c-b6173aa1b161-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e2cad367-9c0f-4556-9e4c-b6173aa1b161\") " pod="openstack/cinder-scheduler-0" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.685852 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2cad367-9c0f-4556-9e4c-b6173aa1b161-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e2cad367-9c0f-4556-9e4c-b6173aa1b161\") " pod="openstack/cinder-scheduler-0" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.685893 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2cad367-9c0f-4556-9e4c-b6173aa1b161-config-data\") pod \"cinder-scheduler-0\" (UID: \"e2cad367-9c0f-4556-9e4c-b6173aa1b161\") " pod="openstack/cinder-scheduler-0" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.685912 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ks82d\" (UniqueName: \"kubernetes.io/projected/e2cad367-9c0f-4556-9e4c-b6173aa1b161-kube-api-access-ks82d\") pod \"cinder-scheduler-0\" (UID: \"e2cad367-9c0f-4556-9e4c-b6173aa1b161\") " pod="openstack/cinder-scheduler-0" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.685961 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2cad367-9c0f-4556-9e4c-b6173aa1b161-scripts\") pod \"cinder-scheduler-0\" (UID: \"e2cad367-9c0f-4556-9e4c-b6173aa1b161\") " pod="openstack/cinder-scheduler-0" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.686047 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e2cad367-9c0f-4556-9e4c-b6173aa1b161-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e2cad367-9c0f-4556-9e4c-b6173aa1b161\") " pod="openstack/cinder-scheduler-0" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.715993 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-79bdd687db-2rnkr" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.768638 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-69677499d6-qsdxd"] Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.768840 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-69677499d6-qsdxd" podUID="e1545875-bc39-40cd-a76a-6229690912a2" containerName="barbican-api-log" containerID="cri-o://b0dd7b1cea28102ec7d70d9ec0f5ed7a01e2a0ffe4c718f45dde21ccc0a0cf4e" gracePeriod=30 Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.769188 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-69677499d6-qsdxd" podUID="e1545875-bc39-40cd-a76a-6229690912a2" containerName="barbican-api" containerID="cri-o://fd344086822719f61ad20d2b0c79d5707af9721b8d69451ce2aa6081ba82ae43" gracePeriod=30 Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.787498 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2cad367-9c0f-4556-9e4c-b6173aa1b161-scripts\") pod \"cinder-scheduler-0\" (UID: \"e2cad367-9c0f-4556-9e4c-b6173aa1b161\") " pod="openstack/cinder-scheduler-0" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.787656 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e2cad367-9c0f-4556-9e4c-b6173aa1b161-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e2cad367-9c0f-4556-9e4c-b6173aa1b161\") " pod="openstack/cinder-scheduler-0" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.787703 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2cad367-9c0f-4556-9e4c-b6173aa1b161-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e2cad367-9c0f-4556-9e4c-b6173aa1b161\") " pod="openstack/cinder-scheduler-0" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.787789 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2cad367-9c0f-4556-9e4c-b6173aa1b161-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e2cad367-9c0f-4556-9e4c-b6173aa1b161\") " pod="openstack/cinder-scheduler-0" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.787826 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2cad367-9c0f-4556-9e4c-b6173aa1b161-config-data\") pod \"cinder-scheduler-0\" (UID: \"e2cad367-9c0f-4556-9e4c-b6173aa1b161\") " pod="openstack/cinder-scheduler-0" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.787842 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ks82d\" (UniqueName: \"kubernetes.io/projected/e2cad367-9c0f-4556-9e4c-b6173aa1b161-kube-api-access-ks82d\") pod \"cinder-scheduler-0\" (UID: \"e2cad367-9c0f-4556-9e4c-b6173aa1b161\") " pod="openstack/cinder-scheduler-0" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.789246 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e2cad367-9c0f-4556-9e4c-b6173aa1b161-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"e2cad367-9c0f-4556-9e4c-b6173aa1b161\") " pod="openstack/cinder-scheduler-0" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.798932 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2cad367-9c0f-4556-9e4c-b6173aa1b161-scripts\") pod \"cinder-scheduler-0\" (UID: \"e2cad367-9c0f-4556-9e4c-b6173aa1b161\") " pod="openstack/cinder-scheduler-0" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.800278 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2cad367-9c0f-4556-9e4c-b6173aa1b161-config-data\") pod \"cinder-scheduler-0\" (UID: \"e2cad367-9c0f-4556-9e4c-b6173aa1b161\") " pod="openstack/cinder-scheduler-0" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.820100 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ks82d\" (UniqueName: \"kubernetes.io/projected/e2cad367-9c0f-4556-9e4c-b6173aa1b161-kube-api-access-ks82d\") pod \"cinder-scheduler-0\" (UID: \"e2cad367-9c0f-4556-9e4c-b6173aa1b161\") " pod="openstack/cinder-scheduler-0" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.824305 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e2cad367-9c0f-4556-9e4c-b6173aa1b161-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"e2cad367-9c0f-4556-9e4c-b6173aa1b161\") " pod="openstack/cinder-scheduler-0" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.849128 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2cad367-9c0f-4556-9e4c-b6173aa1b161-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"e2cad367-9c0f-4556-9e4c-b6173aa1b161\") " pod="openstack/cinder-scheduler-0" Mar 17 09:32:01 crc kubenswrapper[4813]: I0317 09:32:01.964032 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Mar 17 09:32:02 crc kubenswrapper[4813]: I0317 09:32:02.437020 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Mar 17 09:32:02 crc kubenswrapper[4813]: W0317 09:32:02.445927 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode2cad367_9c0f_4556_9e4c_b6173aa1b161.slice/crio-77fa7ebc4bf3acf7e6f412fe5eec7493773c89a513f48ae76c8305bd5685a53d WatchSource:0}: Error finding container 77fa7ebc4bf3acf7e6f412fe5eec7493773c89a513f48ae76c8305bd5685a53d: Status 404 returned error can't find the container with id 77fa7ebc4bf3acf7e6f412fe5eec7493773c89a513f48ae76c8305bd5685a53d Mar 17 09:32:02 crc kubenswrapper[4813]: I0317 09:32:02.565702 4813 generic.go:334] "Generic (PLEG): container finished" podID="e1545875-bc39-40cd-a76a-6229690912a2" containerID="b0dd7b1cea28102ec7d70d9ec0f5ed7a01e2a0ffe4c718f45dde21ccc0a0cf4e" exitCode=143 Mar 17 09:32:02 crc kubenswrapper[4813]: I0317 09:32:02.565957 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-69677499d6-qsdxd" event={"ID":"e1545875-bc39-40cd-a76a-6229690912a2","Type":"ContainerDied","Data":"b0dd7b1cea28102ec7d70d9ec0f5ed7a01e2a0ffe4c718f45dde21ccc0a0cf4e"} Mar 17 09:32:02 crc kubenswrapper[4813]: I0317 09:32:02.567525 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562332-p5phc" event={"ID":"4cf3e585-9c98-4e66-b577-59ddd1327ca2","Type":"ContainerStarted","Data":"dcdf36ad4ecc1c39dbd97f7dff700cc7e19e7a9ed7577401ceb8bb9ee4a3e43d"} Mar 17 09:32:02 crc kubenswrapper[4813]: I0317 09:32:02.571254 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2cad367-9c0f-4556-9e4c-b6173aa1b161","Type":"ContainerStarted","Data":"77fa7ebc4bf3acf7e6f412fe5eec7493773c89a513f48ae76c8305bd5685a53d"} Mar 17 09:32:02 crc kubenswrapper[4813]: I0317 09:32:02.587106 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562332-p5phc" podStartSLOduration=1.5431574179999998 podStartE2EDuration="2.58707567s" podCreationTimestamp="2026-03-17 09:32:00 +0000 UTC" firstStartedPulling="2026-03-17 09:32:01.091823073 +0000 UTC m=+1343.192626572" lastFinishedPulling="2026-03-17 09:32:02.135741325 +0000 UTC m=+1344.236544824" observedRunningTime="2026-03-17 09:32:02.586718999 +0000 UTC m=+1344.687522498" watchObservedRunningTime="2026-03-17 09:32:02.58707567 +0000 UTC m=+1344.687879169" Mar 17 09:32:02 crc kubenswrapper[4813]: I0317 09:32:02.741211 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a5534d6-3502-4a49-8991-cf7fe05d2e8d" path="/var/lib/kubelet/pods/3a5534d6-3502-4a49-8991-cf7fe05d2e8d/volumes" Mar 17 09:32:03 crc kubenswrapper[4813]: I0317 09:32:03.623218 4813 generic.go:334] "Generic (PLEG): container finished" podID="4cf3e585-9c98-4e66-b577-59ddd1327ca2" containerID="dcdf36ad4ecc1c39dbd97f7dff700cc7e19e7a9ed7577401ceb8bb9ee4a3e43d" exitCode=0 Mar 17 09:32:03 crc kubenswrapper[4813]: I0317 09:32:03.623747 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562332-p5phc" event={"ID":"4cf3e585-9c98-4e66-b577-59ddd1327ca2","Type":"ContainerDied","Data":"dcdf36ad4ecc1c39dbd97f7dff700cc7e19e7a9ed7577401ceb8bb9ee4a3e43d"} Mar 17 09:32:03 crc kubenswrapper[4813]: I0317 09:32:03.630927 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2cad367-9c0f-4556-9e4c-b6173aa1b161","Type":"ContainerStarted","Data":"fcd9598e3334758124df28ebe01a621c9d163ddead8b72c1f661802218723ee9"} Mar 17 09:32:03 crc kubenswrapper[4813]: I0317 09:32:03.899049 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Mar 17 09:32:04 crc kubenswrapper[4813]: I0317 09:32:04.662162 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"e2cad367-9c0f-4556-9e4c-b6173aa1b161","Type":"ContainerStarted","Data":"72dadcddc69f9065df5ec29e253f29323691ec76ea78488b0bd0b7217a4d7224"} Mar 17 09:32:04 crc kubenswrapper[4813]: I0317 09:32:04.680762 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.680744567 podStartE2EDuration="3.680744567s" podCreationTimestamp="2026-03-17 09:32:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:32:04.679520478 +0000 UTC m=+1346.780323977" watchObservedRunningTime="2026-03-17 09:32:04.680744567 +0000 UTC m=+1346.781548066" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.083290 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562332-p5phc" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.176427 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5bnl\" (UniqueName: \"kubernetes.io/projected/4cf3e585-9c98-4e66-b577-59ddd1327ca2-kube-api-access-t5bnl\") pod \"4cf3e585-9c98-4e66-b577-59ddd1327ca2\" (UID: \"4cf3e585-9c98-4e66-b577-59ddd1327ca2\") " Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.190768 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cf3e585-9c98-4e66-b577-59ddd1327ca2-kube-api-access-t5bnl" (OuterVolumeSpecName: "kube-api-access-t5bnl") pod "4cf3e585-9c98-4e66-b577-59ddd1327ca2" (UID: "4cf3e585-9c98-4e66-b577-59ddd1327ca2"). InnerVolumeSpecName "kube-api-access-t5bnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.278753 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5bnl\" (UniqueName: \"kubernetes.io/projected/4cf3e585-9c98-4e66-b577-59ddd1327ca2-kube-api-access-t5bnl\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.359819 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.450783 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-8557f6579f-shvsw"] Mar 17 09:32:05 crc kubenswrapper[4813]: E0317 09:32:05.451351 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1545875-bc39-40cd-a76a-6229690912a2" containerName="barbican-api" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.451369 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1545875-bc39-40cd-a76a-6229690912a2" containerName="barbican-api" Mar 17 09:32:05 crc kubenswrapper[4813]: E0317 09:32:05.451387 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1545875-bc39-40cd-a76a-6229690912a2" containerName="barbican-api-log" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.451394 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1545875-bc39-40cd-a76a-6229690912a2" containerName="barbican-api-log" Mar 17 09:32:05 crc kubenswrapper[4813]: E0317 09:32:05.451418 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cf3e585-9c98-4e66-b577-59ddd1327ca2" containerName="oc" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.451424 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cf3e585-9c98-4e66-b577-59ddd1327ca2" containerName="oc" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.451672 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cf3e585-9c98-4e66-b577-59ddd1327ca2" containerName="oc" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.451707 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1545875-bc39-40cd-a76a-6229690912a2" containerName="barbican-api-log" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.451727 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1545875-bc39-40cd-a76a-6229690912a2" containerName="barbican-api" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.457426 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.461665 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.461740 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.461817 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.466901 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-8557f6579f-shvsw"] Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.482864 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1545875-bc39-40cd-a76a-6229690912a2-config-data\") pod \"e1545875-bc39-40cd-a76a-6229690912a2\" (UID: \"e1545875-bc39-40cd-a76a-6229690912a2\") " Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.482924 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1545875-bc39-40cd-a76a-6229690912a2-combined-ca-bundle\") pod \"e1545875-bc39-40cd-a76a-6229690912a2\" (UID: \"e1545875-bc39-40cd-a76a-6229690912a2\") " Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.483047 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1545875-bc39-40cd-a76a-6229690912a2-logs\") pod \"e1545875-bc39-40cd-a76a-6229690912a2\" (UID: \"e1545875-bc39-40cd-a76a-6229690912a2\") " Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.483109 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z448k\" (UniqueName: \"kubernetes.io/projected/e1545875-bc39-40cd-a76a-6229690912a2-kube-api-access-z448k\") pod \"e1545875-bc39-40cd-a76a-6229690912a2\" (UID: \"e1545875-bc39-40cd-a76a-6229690912a2\") " Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.483161 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e1545875-bc39-40cd-a76a-6229690912a2-config-data-custom\") pod \"e1545875-bc39-40cd-a76a-6229690912a2\" (UID: \"e1545875-bc39-40cd-a76a-6229690912a2\") " Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.485455 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e1545875-bc39-40cd-a76a-6229690912a2-logs" (OuterVolumeSpecName: "logs") pod "e1545875-bc39-40cd-a76a-6229690912a2" (UID: "e1545875-bc39-40cd-a76a-6229690912a2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.490175 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1545875-bc39-40cd-a76a-6229690912a2-kube-api-access-z448k" (OuterVolumeSpecName: "kube-api-access-z448k") pod "e1545875-bc39-40cd-a76a-6229690912a2" (UID: "e1545875-bc39-40cd-a76a-6229690912a2"). InnerVolumeSpecName "kube-api-access-z448k". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.493752 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1545875-bc39-40cd-a76a-6229690912a2-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e1545875-bc39-40cd-a76a-6229690912a2" (UID: "e1545875-bc39-40cd-a76a-6229690912a2"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.530673 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1545875-bc39-40cd-a76a-6229690912a2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e1545875-bc39-40cd-a76a-6229690912a2" (UID: "e1545875-bc39-40cd-a76a-6229690912a2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.539726 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1545875-bc39-40cd-a76a-6229690912a2-config-data" (OuterVolumeSpecName: "config-data") pod "e1545875-bc39-40cd-a76a-6229690912a2" (UID: "e1545875-bc39-40cd-a76a-6229690912a2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.585390 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e79b20b0-682e-4d16-bb56-64f0c4ec0202-combined-ca-bundle\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.585712 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vm9pl\" (UniqueName: \"kubernetes.io/projected/e79b20b0-682e-4d16-bb56-64f0c4ec0202-kube-api-access-vm9pl\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.585825 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e79b20b0-682e-4d16-bb56-64f0c4ec0202-public-tls-certs\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.585934 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e79b20b0-682e-4d16-bb56-64f0c4ec0202-log-httpd\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.586065 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e79b20b0-682e-4d16-bb56-64f0c4ec0202-config-data\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.586185 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e79b20b0-682e-4d16-bb56-64f0c4ec0202-etc-swift\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.586286 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e79b20b0-682e-4d16-bb56-64f0c4ec0202-internal-tls-certs\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.586437 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e79b20b0-682e-4d16-bb56-64f0c4ec0202-run-httpd\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.586638 4813 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e1545875-bc39-40cd-a76a-6229690912a2-config-data-custom\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.586723 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e1545875-bc39-40cd-a76a-6229690912a2-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.586798 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e1545875-bc39-40cd-a76a-6229690912a2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.586872 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e1545875-bc39-40cd-a76a-6229690912a2-logs\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.586930 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z448k\" (UniqueName: \"kubernetes.io/projected/e1545875-bc39-40cd-a76a-6229690912a2-kube-api-access-z448k\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.641232 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562326-rbvpg"] Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.649816 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562326-rbvpg"] Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.656642 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.657998 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.664072 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.664674 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-vmn85" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.664724 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.664872 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.679857 4813 generic.go:334] "Generic (PLEG): container finished" podID="e1545875-bc39-40cd-a76a-6229690912a2" containerID="fd344086822719f61ad20d2b0c79d5707af9721b8d69451ce2aa6081ba82ae43" exitCode=0 Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.679922 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-69677499d6-qsdxd" event={"ID":"e1545875-bc39-40cd-a76a-6229690912a2","Type":"ContainerDied","Data":"fd344086822719f61ad20d2b0c79d5707af9721b8d69451ce2aa6081ba82ae43"} Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.679948 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-69677499d6-qsdxd" event={"ID":"e1545875-bc39-40cd-a76a-6229690912a2","Type":"ContainerDied","Data":"5429603a853162372b069be1edffa26fc59fd14f2b8b7a6e6c6401e5f124505b"} Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.679964 4813 scope.go:117] "RemoveContainer" containerID="fd344086822719f61ad20d2b0c79d5707af9721b8d69451ce2aa6081ba82ae43" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.680056 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-69677499d6-qsdxd" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.689274 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e79b20b0-682e-4d16-bb56-64f0c4ec0202-run-httpd\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.689302 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562332-p5phc" event={"ID":"4cf3e585-9c98-4e66-b577-59ddd1327ca2","Type":"ContainerDied","Data":"6d819e4356134cbb82a3b591162f32a5f6122f65ebe6c008d14c32282e409b79"} Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.689335 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6d819e4356134cbb82a3b591162f32a5f6122f65ebe6c008d14c32282e409b79" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.689356 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e79b20b0-682e-4d16-bb56-64f0c4ec0202-combined-ca-bundle\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.689385 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vm9pl\" (UniqueName: \"kubernetes.io/projected/e79b20b0-682e-4d16-bb56-64f0c4ec0202-kube-api-access-vm9pl\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.689406 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e79b20b0-682e-4d16-bb56-64f0c4ec0202-public-tls-certs\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.689434 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e79b20b0-682e-4d16-bb56-64f0c4ec0202-log-httpd\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.689461 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e79b20b0-682e-4d16-bb56-64f0c4ec0202-config-data\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.689491 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e79b20b0-682e-4d16-bb56-64f0c4ec0202-etc-swift\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.689507 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e79b20b0-682e-4d16-bb56-64f0c4ec0202-internal-tls-certs\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.690189 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562332-p5phc" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.690404 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e79b20b0-682e-4d16-bb56-64f0c4ec0202-run-httpd\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.690475 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e79b20b0-682e-4d16-bb56-64f0c4ec0202-log-httpd\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.694242 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e79b20b0-682e-4d16-bb56-64f0c4ec0202-combined-ca-bundle\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.694619 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e79b20b0-682e-4d16-bb56-64f0c4ec0202-config-data\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.695221 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/e79b20b0-682e-4d16-bb56-64f0c4ec0202-etc-swift\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.696374 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e79b20b0-682e-4d16-bb56-64f0c4ec0202-internal-tls-certs\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.702185 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e79b20b0-682e-4d16-bb56-64f0c4ec0202-public-tls-certs\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.709912 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vm9pl\" (UniqueName: \"kubernetes.io/projected/e79b20b0-682e-4d16-bb56-64f0c4ec0202-kube-api-access-vm9pl\") pod \"swift-proxy-8557f6579f-shvsw\" (UID: \"e79b20b0-682e-4d16-bb56-64f0c4ec0202\") " pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.713641 4813 scope.go:117] "RemoveContainer" containerID="b0dd7b1cea28102ec7d70d9ec0f5ed7a01e2a0ffe4c718f45dde21ccc0a0cf4e" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.721005 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-69677499d6-qsdxd"] Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.735294 4813 scope.go:117] "RemoveContainer" containerID="fd344086822719f61ad20d2b0c79d5707af9721b8d69451ce2aa6081ba82ae43" Mar 17 09:32:05 crc kubenswrapper[4813]: E0317 09:32:05.735700 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd344086822719f61ad20d2b0c79d5707af9721b8d69451ce2aa6081ba82ae43\": container with ID starting with fd344086822719f61ad20d2b0c79d5707af9721b8d69451ce2aa6081ba82ae43 not found: ID does not exist" containerID="fd344086822719f61ad20d2b0c79d5707af9721b8d69451ce2aa6081ba82ae43" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.735747 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd344086822719f61ad20d2b0c79d5707af9721b8d69451ce2aa6081ba82ae43"} err="failed to get container status \"fd344086822719f61ad20d2b0c79d5707af9721b8d69451ce2aa6081ba82ae43\": rpc error: code = NotFound desc = could not find container \"fd344086822719f61ad20d2b0c79d5707af9721b8d69451ce2aa6081ba82ae43\": container with ID starting with fd344086822719f61ad20d2b0c79d5707af9721b8d69451ce2aa6081ba82ae43 not found: ID does not exist" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.735773 4813 scope.go:117] "RemoveContainer" containerID="b0dd7b1cea28102ec7d70d9ec0f5ed7a01e2a0ffe4c718f45dde21ccc0a0cf4e" Mar 17 09:32:05 crc kubenswrapper[4813]: E0317 09:32:05.736340 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0dd7b1cea28102ec7d70d9ec0f5ed7a01e2a0ffe4c718f45dde21ccc0a0cf4e\": container with ID starting with b0dd7b1cea28102ec7d70d9ec0f5ed7a01e2a0ffe4c718f45dde21ccc0a0cf4e not found: ID does not exist" containerID="b0dd7b1cea28102ec7d70d9ec0f5ed7a01e2a0ffe4c718f45dde21ccc0a0cf4e" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.736371 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0dd7b1cea28102ec7d70d9ec0f5ed7a01e2a0ffe4c718f45dde21ccc0a0cf4e"} err="failed to get container status \"b0dd7b1cea28102ec7d70d9ec0f5ed7a01e2a0ffe4c718f45dde21ccc0a0cf4e\": rpc error: code = NotFound desc = could not find container \"b0dd7b1cea28102ec7d70d9ec0f5ed7a01e2a0ffe4c718f45dde21ccc0a0cf4e\": container with ID starting with b0dd7b1cea28102ec7d70d9ec0f5ed7a01e2a0ffe4c718f45dde21ccc0a0cf4e not found: ID does not exist" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.739926 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-69677499d6-qsdxd"] Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.782003 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.791048 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c532fc6b-8310-4322-b09c-bd21b19154ec-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c532fc6b-8310-4322-b09c-bd21b19154ec\") " pod="openstack/openstackclient" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.791477 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjjf2\" (UniqueName: \"kubernetes.io/projected/c532fc6b-8310-4322-b09c-bd21b19154ec-kube-api-access-rjjf2\") pod \"openstackclient\" (UID: \"c532fc6b-8310-4322-b09c-bd21b19154ec\") " pod="openstack/openstackclient" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.791522 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c532fc6b-8310-4322-b09c-bd21b19154ec-openstack-config\") pod \"openstackclient\" (UID: \"c532fc6b-8310-4322-b09c-bd21b19154ec\") " pod="openstack/openstackclient" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.791747 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c532fc6b-8310-4322-b09c-bd21b19154ec-openstack-config-secret\") pod \"openstackclient\" (UID: \"c532fc6b-8310-4322-b09c-bd21b19154ec\") " pod="openstack/openstackclient" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.894567 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjjf2\" (UniqueName: \"kubernetes.io/projected/c532fc6b-8310-4322-b09c-bd21b19154ec-kube-api-access-rjjf2\") pod \"openstackclient\" (UID: \"c532fc6b-8310-4322-b09c-bd21b19154ec\") " pod="openstack/openstackclient" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.894618 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c532fc6b-8310-4322-b09c-bd21b19154ec-openstack-config\") pod \"openstackclient\" (UID: \"c532fc6b-8310-4322-b09c-bd21b19154ec\") " pod="openstack/openstackclient" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.894658 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c532fc6b-8310-4322-b09c-bd21b19154ec-openstack-config-secret\") pod \"openstackclient\" (UID: \"c532fc6b-8310-4322-b09c-bd21b19154ec\") " pod="openstack/openstackclient" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.894701 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c532fc6b-8310-4322-b09c-bd21b19154ec-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c532fc6b-8310-4322-b09c-bd21b19154ec\") " pod="openstack/openstackclient" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.900754 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c532fc6b-8310-4322-b09c-bd21b19154ec-openstack-config\") pod \"openstackclient\" (UID: \"c532fc6b-8310-4322-b09c-bd21b19154ec\") " pod="openstack/openstackclient" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.906428 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c532fc6b-8310-4322-b09c-bd21b19154ec-combined-ca-bundle\") pod \"openstackclient\" (UID: \"c532fc6b-8310-4322-b09c-bd21b19154ec\") " pod="openstack/openstackclient" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.912559 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c532fc6b-8310-4322-b09c-bd21b19154ec-openstack-config-secret\") pod \"openstackclient\" (UID: \"c532fc6b-8310-4322-b09c-bd21b19154ec\") " pod="openstack/openstackclient" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.949169 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjjf2\" (UniqueName: \"kubernetes.io/projected/c532fc6b-8310-4322-b09c-bd21b19154ec-kube-api-access-rjjf2\") pod \"openstackclient\" (UID: \"c532fc6b-8310-4322-b09c-bd21b19154ec\") " pod="openstack/openstackclient" Mar 17 09:32:05 crc kubenswrapper[4813]: I0317 09:32:05.981038 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Mar 17 09:32:06 crc kubenswrapper[4813]: I0317 09:32:06.413364 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-8557f6579f-shvsw"] Mar 17 09:32:06 crc kubenswrapper[4813]: I0317 09:32:06.525385 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Mar 17 09:32:06 crc kubenswrapper[4813]: I0317 09:32:06.699310 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-8557f6579f-shvsw" event={"ID":"e79b20b0-682e-4d16-bb56-64f0c4ec0202","Type":"ContainerStarted","Data":"6337990fcec28c9a8e7a364f24b2fe26e7fb0412fcd9211e5d507f94440763c6"} Mar 17 09:32:06 crc kubenswrapper[4813]: I0317 09:32:06.699358 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-8557f6579f-shvsw" event={"ID":"e79b20b0-682e-4d16-bb56-64f0c4ec0202","Type":"ContainerStarted","Data":"2e07bb5bbee425bd592347613377743e4c521949e2d77f61906d19d5b93ca659"} Mar 17 09:32:06 crc kubenswrapper[4813]: I0317 09:32:06.701014 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"c532fc6b-8310-4322-b09c-bd21b19154ec","Type":"ContainerStarted","Data":"5a9acc9cfca944f2d5eecba24f8a27716df82dc21a35fa9ec383cb677b5455fb"} Mar 17 09:32:06 crc kubenswrapper[4813]: I0317 09:32:06.742124 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d5793f57-43fd-4c88-b61a-50702ff099ce" path="/var/lib/kubelet/pods/d5793f57-43fd-4c88-b61a-50702ff099ce/volumes" Mar 17 09:32:06 crc kubenswrapper[4813]: I0317 09:32:06.743052 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1545875-bc39-40cd-a76a-6229690912a2" path="/var/lib/kubelet/pods/e1545875-bc39-40cd-a76a-6229690912a2/volumes" Mar 17 09:32:06 crc kubenswrapper[4813]: I0317 09:32:06.964899 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Mar 17 09:32:07 crc kubenswrapper[4813]: I0317 09:32:07.721163 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-8557f6579f-shvsw" event={"ID":"e79b20b0-682e-4d16-bb56-64f0c4ec0202","Type":"ContainerStarted","Data":"622dea3bcf098a602ec97188a3eb08ead6323ef35491f1062143468a8197ffa8"} Mar 17 09:32:07 crc kubenswrapper[4813]: I0317 09:32:07.721466 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:07 crc kubenswrapper[4813]: I0317 09:32:07.721510 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:07 crc kubenswrapper[4813]: I0317 09:32:07.749029 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-8557f6579f-shvsw" podStartSLOduration=2.749008017 podStartE2EDuration="2.749008017s" podCreationTimestamp="2026-03-17 09:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:32:07.743188913 +0000 UTC m=+1349.843992412" watchObservedRunningTime="2026-03-17 09:32:07.749008017 +0000 UTC m=+1349.849811516" Mar 17 09:32:10 crc kubenswrapper[4813]: I0317 09:32:10.306091 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-85fc85fdfb-n469t" podUID="0a52b24b-cc19-4720-a05b-961015362769" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.156:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.156:8443: connect: connection refused" Mar 17 09:32:12 crc kubenswrapper[4813]: I0317 09:32:12.181075 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Mar 17 09:32:12 crc kubenswrapper[4813]: I0317 09:32:12.354213 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 09:32:12 crc kubenswrapper[4813]: I0317 09:32:12.354501 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a1c5331f-3325-4bab-a0de-016366febfcd" containerName="glance-log" containerID="cri-o://b571f7ac595fd0a9269bdb93504580575de28b2f8b55d4c8d29eb60d02fad58e" gracePeriod=30 Mar 17 09:32:12 crc kubenswrapper[4813]: I0317 09:32:12.354696 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="a1c5331f-3325-4bab-a0de-016366febfcd" containerName="glance-httpd" containerID="cri-o://6a8330860efcd7dff1fc03e67f918941c9765c0844a3b969348cd2ca24a98307" gracePeriod=30 Mar 17 09:32:12 crc kubenswrapper[4813]: I0317 09:32:12.786238 4813 generic.go:334] "Generic (PLEG): container finished" podID="a1c5331f-3325-4bab-a0de-016366febfcd" containerID="b571f7ac595fd0a9269bdb93504580575de28b2f8b55d4c8d29eb60d02fad58e" exitCode=143 Mar 17 09:32:12 crc kubenswrapper[4813]: I0317 09:32:12.786414 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a1c5331f-3325-4bab-a0de-016366febfcd","Type":"ContainerDied","Data":"b571f7ac595fd0a9269bdb93504580575de28b2f8b55d4c8d29eb60d02fad58e"} Mar 17 09:32:15 crc kubenswrapper[4813]: I0317 09:32:15.789846 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:15 crc kubenswrapper[4813]: I0317 09:32:15.793677 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-8557f6579f-shvsw" Mar 17 09:32:15 crc kubenswrapper[4813]: I0317 09:32:15.805112 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 09:32:15 crc kubenswrapper[4813]: I0317 09:32:15.805343 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="fc74e31f-b87b-4888-b99f-342c2f72f6b7" containerName="glance-log" containerID="cri-o://20afc3b693977dd502ec4c0428f5b9fd7987ff5309f74555d32af8f5c64e9012" gracePeriod=30 Mar 17 09:32:15 crc kubenswrapper[4813]: I0317 09:32:15.807425 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="fc74e31f-b87b-4888-b99f-342c2f72f6b7" containerName="glance-httpd" containerID="cri-o://12e6c7f140dd842351032cd69a0fd18874b77ca813d3ea3f119f9553b1b4b862" gracePeriod=30 Mar 17 09:32:15 crc kubenswrapper[4813]: I0317 09:32:15.840923 4813 generic.go:334] "Generic (PLEG): container finished" podID="a1c5331f-3325-4bab-a0de-016366febfcd" containerID="6a8330860efcd7dff1fc03e67f918941c9765c0844a3b969348cd2ca24a98307" exitCode=0 Mar 17 09:32:15 crc kubenswrapper[4813]: I0317 09:32:15.840998 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a1c5331f-3325-4bab-a0de-016366febfcd","Type":"ContainerDied","Data":"6a8330860efcd7dff1fc03e67f918941c9765c0844a3b969348cd2ca24a98307"} Mar 17 09:32:15 crc kubenswrapper[4813]: I0317 09:32:15.853931 4813 generic.go:334] "Generic (PLEG): container finished" podID="d927c167-348f-47f8-9860-b21dac3ca309" containerID="db2d07ad344f758a6f4de8415ae1005b4319f5b36eae2e9f1be1bbb8df4134d3" exitCode=137 Mar 17 09:32:15 crc kubenswrapper[4813]: I0317 09:32:15.854207 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d927c167-348f-47f8-9860-b21dac3ca309","Type":"ContainerDied","Data":"db2d07ad344f758a6f4de8415ae1005b4319f5b36eae2e9f1be1bbb8df4134d3"} Mar 17 09:32:15 crc kubenswrapper[4813]: I0317 09:32:15.891101 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.009017 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-config-data\") pod \"d927c167-348f-47f8-9860-b21dac3ca309\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.009088 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-combined-ca-bundle\") pod \"d927c167-348f-47f8-9860-b21dac3ca309\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.009115 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d927c167-348f-47f8-9860-b21dac3ca309-log-httpd\") pod \"d927c167-348f-47f8-9860-b21dac3ca309\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.009134 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4c7k8\" (UniqueName: \"kubernetes.io/projected/d927c167-348f-47f8-9860-b21dac3ca309-kube-api-access-4c7k8\") pod \"d927c167-348f-47f8-9860-b21dac3ca309\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.009156 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d927c167-348f-47f8-9860-b21dac3ca309-run-httpd\") pod \"d927c167-348f-47f8-9860-b21dac3ca309\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.009206 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-scripts\") pod \"d927c167-348f-47f8-9860-b21dac3ca309\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.009252 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-sg-core-conf-yaml\") pod \"d927c167-348f-47f8-9860-b21dac3ca309\" (UID: \"d927c167-348f-47f8-9860-b21dac3ca309\") " Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.009496 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d927c167-348f-47f8-9860-b21dac3ca309-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d927c167-348f-47f8-9860-b21dac3ca309" (UID: "d927c167-348f-47f8-9860-b21dac3ca309"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.009803 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d927c167-348f-47f8-9860-b21dac3ca309-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d927c167-348f-47f8-9860-b21dac3ca309" (UID: "d927c167-348f-47f8-9860-b21dac3ca309"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.010065 4813 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d927c167-348f-47f8-9860-b21dac3ca309-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.010078 4813 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d927c167-348f-47f8-9860-b21dac3ca309-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.020253 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-scripts" (OuterVolumeSpecName: "scripts") pod "d927c167-348f-47f8-9860-b21dac3ca309" (UID: "d927c167-348f-47f8-9860-b21dac3ca309"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.029701 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d927c167-348f-47f8-9860-b21dac3ca309-kube-api-access-4c7k8" (OuterVolumeSpecName: "kube-api-access-4c7k8") pod "d927c167-348f-47f8-9860-b21dac3ca309" (UID: "d927c167-348f-47f8-9860-b21dac3ca309"). InnerVolumeSpecName "kube-api-access-4c7k8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.055723 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d927c167-348f-47f8-9860-b21dac3ca309" (UID: "d927c167-348f-47f8-9860-b21dac3ca309"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.101458 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d927c167-348f-47f8-9860-b21dac3ca309" (UID: "d927c167-348f-47f8-9860-b21dac3ca309"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.111810 4813 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.111835 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.111845 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4c7k8\" (UniqueName: \"kubernetes.io/projected/d927c167-348f-47f8-9860-b21dac3ca309-kube-api-access-4c7k8\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.111856 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.128448 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-config-data" (OuterVolumeSpecName: "config-data") pod "d927c167-348f-47f8-9860-b21dac3ca309" (UID: "d927c167-348f-47f8-9860-b21dac3ca309"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.131128 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.213977 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52qfr\" (UniqueName: \"kubernetes.io/projected/a1c5331f-3325-4bab-a0de-016366febfcd-kube-api-access-52qfr\") pod \"a1c5331f-3325-4bab-a0de-016366febfcd\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.214291 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-config-data\") pod \"a1c5331f-3325-4bab-a0de-016366febfcd\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.214391 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c5331f-3325-4bab-a0de-016366febfcd-logs\") pod \"a1c5331f-3325-4bab-a0de-016366febfcd\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.214407 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-public-tls-certs\") pod \"a1c5331f-3325-4bab-a0de-016366febfcd\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.214434 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-combined-ca-bundle\") pod \"a1c5331f-3325-4bab-a0de-016366febfcd\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.214501 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1c5331f-3325-4bab-a0de-016366febfcd-httpd-run\") pod \"a1c5331f-3325-4bab-a0de-016366febfcd\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.214538 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-scripts\") pod \"a1c5331f-3325-4bab-a0de-016366febfcd\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.214565 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"a1c5331f-3325-4bab-a0de-016366febfcd\" (UID: \"a1c5331f-3325-4bab-a0de-016366febfcd\") " Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.214939 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d927c167-348f-47f8-9860-b21dac3ca309-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.215793 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1c5331f-3325-4bab-a0de-016366febfcd-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a1c5331f-3325-4bab-a0de-016366febfcd" (UID: "a1c5331f-3325-4bab-a0de-016366febfcd"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.220706 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-scripts" (OuterVolumeSpecName: "scripts") pod "a1c5331f-3325-4bab-a0de-016366febfcd" (UID: "a1c5331f-3325-4bab-a0de-016366febfcd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.220700 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1c5331f-3325-4bab-a0de-016366febfcd-logs" (OuterVolumeSpecName: "logs") pod "a1c5331f-3325-4bab-a0de-016366febfcd" (UID: "a1c5331f-3325-4bab-a0de-016366febfcd"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.221260 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance") pod "a1c5331f-3325-4bab-a0de-016366febfcd" (UID: "a1c5331f-3325-4bab-a0de-016366febfcd"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.223715 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1c5331f-3325-4bab-a0de-016366febfcd-kube-api-access-52qfr" (OuterVolumeSpecName: "kube-api-access-52qfr") pod "a1c5331f-3325-4bab-a0de-016366febfcd" (UID: "a1c5331f-3325-4bab-a0de-016366febfcd"). InnerVolumeSpecName "kube-api-access-52qfr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.240316 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a1c5331f-3325-4bab-a0de-016366febfcd" (UID: "a1c5331f-3325-4bab-a0de-016366febfcd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.258093 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-config-data" (OuterVolumeSpecName: "config-data") pod "a1c5331f-3325-4bab-a0de-016366febfcd" (UID: "a1c5331f-3325-4bab-a0de-016366febfcd"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.290796 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "a1c5331f-3325-4bab-a0de-016366febfcd" (UID: "a1c5331f-3325-4bab-a0de-016366febfcd"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.316329 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c5331f-3325-4bab-a0de-016366febfcd-logs\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.316364 4813 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.316375 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.316383 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a1c5331f-3325-4bab-a0de-016366febfcd-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.316391 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.316425 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.316434 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52qfr\" (UniqueName: \"kubernetes.io/projected/a1c5331f-3325-4bab-a0de-016366febfcd-kube-api-access-52qfr\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.316447 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1c5331f-3325-4bab-a0de-016366febfcd-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.334216 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.418412 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.862273 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"c532fc6b-8310-4322-b09c-bd21b19154ec","Type":"ContainerStarted","Data":"8892dda9b33bbade98b91fcfcf02effc27e22a7eafc77a9300f6097ae98e8487"} Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.865493 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"a1c5331f-3325-4bab-a0de-016366febfcd","Type":"ContainerDied","Data":"6b71041d6c374849f3739c6a41ddf51e58df25ee101febcff6c3077ccad61165"} Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.865538 4813 scope.go:117] "RemoveContainer" containerID="6a8330860efcd7dff1fc03e67f918941c9765c0844a3b969348cd2ca24a98307" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.865650 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.871890 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d927c167-348f-47f8-9860-b21dac3ca309","Type":"ContainerDied","Data":"9e18c6e47f25e05f0bc375bef10326d7fd8e3b51c7be154866b1604721bfb3da"} Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.872086 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.876877 4813 generic.go:334] "Generic (PLEG): container finished" podID="fc74e31f-b87b-4888-b99f-342c2f72f6b7" containerID="20afc3b693977dd502ec4c0428f5b9fd7987ff5309f74555d32af8f5c64e9012" exitCode=143 Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.876911 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fc74e31f-b87b-4888-b99f-342c2f72f6b7","Type":"ContainerDied","Data":"20afc3b693977dd502ec4c0428f5b9fd7987ff5309f74555d32af8f5c64e9012"} Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.885447 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.75323459 podStartE2EDuration="11.885434405s" podCreationTimestamp="2026-03-17 09:32:05 +0000 UTC" firstStartedPulling="2026-03-17 09:32:06.533214029 +0000 UTC m=+1348.634017528" lastFinishedPulling="2026-03-17 09:32:15.665413844 +0000 UTC m=+1357.766217343" observedRunningTime="2026-03-17 09:32:16.884995441 +0000 UTC m=+1358.985798980" watchObservedRunningTime="2026-03-17 09:32:16.885434405 +0000 UTC m=+1358.986237904" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.895413 4813 scope.go:117] "RemoveContainer" containerID="b571f7ac595fd0a9269bdb93504580575de28b2f8b55d4c8d29eb60d02fad58e" Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.907377 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 09:32:16 crc kubenswrapper[4813]: I0317 09:32:16.945587 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.016246 4813 scope.go:117] "RemoveContainer" containerID="db2d07ad344f758a6f4de8415ae1005b4319f5b36eae2e9f1be1bbb8df4134d3" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.018996 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 09:32:17 crc kubenswrapper[4813]: E0317 09:32:17.019528 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d927c167-348f-47f8-9860-b21dac3ca309" containerName="sg-core" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.019545 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d927c167-348f-47f8-9860-b21dac3ca309" containerName="sg-core" Mar 17 09:32:17 crc kubenswrapper[4813]: E0317 09:32:17.019554 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c5331f-3325-4bab-a0de-016366febfcd" containerName="glance-log" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.019561 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c5331f-3325-4bab-a0de-016366febfcd" containerName="glance-log" Mar 17 09:32:17 crc kubenswrapper[4813]: E0317 09:32:17.019584 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d927c167-348f-47f8-9860-b21dac3ca309" containerName="proxy-httpd" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.019590 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d927c167-348f-47f8-9860-b21dac3ca309" containerName="proxy-httpd" Mar 17 09:32:17 crc kubenswrapper[4813]: E0317 09:32:17.019616 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d927c167-348f-47f8-9860-b21dac3ca309" containerName="ceilometer-notification-agent" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.019622 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d927c167-348f-47f8-9860-b21dac3ca309" containerName="ceilometer-notification-agent" Mar 17 09:32:17 crc kubenswrapper[4813]: E0317 09:32:17.019637 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1c5331f-3325-4bab-a0de-016366febfcd" containerName="glance-httpd" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.019644 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1c5331f-3325-4bab-a0de-016366febfcd" containerName="glance-httpd" Mar 17 09:32:17 crc kubenswrapper[4813]: E0317 09:32:17.019655 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d927c167-348f-47f8-9860-b21dac3ca309" containerName="ceilometer-central-agent" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.019661 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d927c167-348f-47f8-9860-b21dac3ca309" containerName="ceilometer-central-agent" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.019809 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d927c167-348f-47f8-9860-b21dac3ca309" containerName="ceilometer-notification-agent" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.019821 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1c5331f-3325-4bab-a0de-016366febfcd" containerName="glance-log" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.019832 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d927c167-348f-47f8-9860-b21dac3ca309" containerName="ceilometer-central-agent" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.019843 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1c5331f-3325-4bab-a0de-016366febfcd" containerName="glance-httpd" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.019854 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d927c167-348f-47f8-9860-b21dac3ca309" containerName="sg-core" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.019871 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d927c167-348f-47f8-9860-b21dac3ca309" containerName="proxy-httpd" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.020924 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.025612 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.025772 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.028103 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.036996 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.047447 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.055613 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.056952 4813 scope.go:117] "RemoveContainer" containerID="faad4b4a7c4535789ba6e076f9588c4385a53196f95b284dcf17a03fd3bc815f" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.058104 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.060664 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.060705 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.064129 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.086443 4813 scope.go:117] "RemoveContainer" containerID="996ef8dc81ee850c6bcdf1656dc2e58131b9cf47f3b83174cf9f63eaddb3e604" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.110784 4813 scope.go:117] "RemoveContainer" containerID="766a0ef148bc4a1cc73bfe82a1e3d829969f16ee65cb1fd39fd23aac1184be02" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.134977 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b4ddd1-d517-43cc-ac5e-636026efff5d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.135022 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b4ddd1-d517-43cc-ac5e-636026efff5d-config-data\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.135154 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-config-data\") pod \"ceilometer-0\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.135262 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.135335 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-scripts\") pod \"ceilometer-0\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.135438 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f4b4ddd1-d517-43cc-ac5e-636026efff5d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.135490 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4b4ddd1-d517-43cc-ac5e-636026efff5d-scripts\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.135532 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qctnb\" (UniqueName: \"kubernetes.io/projected/f4b4ddd1-d517-43cc-ac5e-636026efff5d-kube-api-access-qctnb\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.135574 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92f6c71b-18e8-4015-a231-a741629a4711-log-httpd\") pod \"ceilometer-0\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.135698 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j75tr\" (UniqueName: \"kubernetes.io/projected/92f6c71b-18e8-4015-a231-a741629a4711-kube-api-access-j75tr\") pod \"ceilometer-0\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.135743 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.135779 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.135830 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4b4ddd1-d517-43cc-ac5e-636026efff5d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.135937 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4b4ddd1-d517-43cc-ac5e-636026efff5d-logs\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.136019 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92f6c71b-18e8-4015-a231-a741629a4711-run-httpd\") pod \"ceilometer-0\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.237792 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92f6c71b-18e8-4015-a231-a741629a4711-run-httpd\") pod \"ceilometer-0\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.237857 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b4ddd1-d517-43cc-ac5e-636026efff5d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.237894 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b4ddd1-d517-43cc-ac5e-636026efff5d-config-data\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.237922 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-config-data\") pod \"ceilometer-0\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.237959 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.237992 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-scripts\") pod \"ceilometer-0\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.238033 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f4b4ddd1-d517-43cc-ac5e-636026efff5d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.238062 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4b4ddd1-d517-43cc-ac5e-636026efff5d-scripts\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.238091 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qctnb\" (UniqueName: \"kubernetes.io/projected/f4b4ddd1-d517-43cc-ac5e-636026efff5d-kube-api-access-qctnb\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.238117 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92f6c71b-18e8-4015-a231-a741629a4711-log-httpd\") pod \"ceilometer-0\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.238313 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.238635 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92f6c71b-18e8-4015-a231-a741629a4711-run-httpd\") pod \"ceilometer-0\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.238694 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f4b4ddd1-d517-43cc-ac5e-636026efff5d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.238704 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92f6c71b-18e8-4015-a231-a741629a4711-log-httpd\") pod \"ceilometer-0\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.238756 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j75tr\" (UniqueName: \"kubernetes.io/projected/92f6c71b-18e8-4015-a231-a741629a4711-kube-api-access-j75tr\") pod \"ceilometer-0\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.238787 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.238818 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.238856 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4b4ddd1-d517-43cc-ac5e-636026efff5d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.239296 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4b4ddd1-d517-43cc-ac5e-636026efff5d-logs\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.239577 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f4b4ddd1-d517-43cc-ac5e-636026efff5d-logs\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.245778 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f4b4ddd1-d517-43cc-ac5e-636026efff5d-config-data\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.247795 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.247856 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f4b4ddd1-d517-43cc-ac5e-636026efff5d-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.248038 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f4b4ddd1-d517-43cc-ac5e-636026efff5d-scripts\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.248827 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.249167 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f4b4ddd1-d517-43cc-ac5e-636026efff5d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.253812 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-config-data\") pod \"ceilometer-0\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.255247 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-scripts\") pod \"ceilometer-0\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.259539 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j75tr\" (UniqueName: \"kubernetes.io/projected/92f6c71b-18e8-4015-a231-a741629a4711-kube-api-access-j75tr\") pod \"ceilometer-0\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.263225 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qctnb\" (UniqueName: \"kubernetes.io/projected/f4b4ddd1-d517-43cc-ac5e-636026efff5d-kube-api-access-qctnb\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.267428 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-external-api-0\" (UID: \"f4b4ddd1-d517-43cc-ac5e-636026efff5d\") " pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.341705 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.384245 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.637922 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.869987 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Mar 17 09:32:17 crc kubenswrapper[4813]: W0317 09:32:17.891258 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf4b4ddd1_d517_43cc_ac5e_636026efff5d.slice/crio-b3184844bed4947d99d54a6d396f850d94c4d8f719f619c1cb685918cea9b715 WatchSource:0}: Error finding container b3184844bed4947d99d54a6d396f850d94c4d8f719f619c1cb685918cea9b715: Status 404 returned error can't find the container with id b3184844bed4947d99d54a6d396f850d94c4d8f719f619c1cb685918cea9b715 Mar 17 09:32:17 crc kubenswrapper[4813]: I0317 09:32:17.969867 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:32:17 crc kubenswrapper[4813]: W0317 09:32:17.978059 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92f6c71b_18e8_4015_a231_a741629a4711.slice/crio-d386ba7487283c309d72f0e6c93fd1891a38efa8f8fc1ffec822391cf60f2d28 WatchSource:0}: Error finding container d386ba7487283c309d72f0e6c93fd1891a38efa8f8fc1ffec822391cf60f2d28: Status 404 returned error can't find the container with id d386ba7487283c309d72f0e6c93fd1891a38efa8f8fc1ffec822391cf60f2d28 Mar 17 09:32:18 crc kubenswrapper[4813]: I0317 09:32:18.751831 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1c5331f-3325-4bab-a0de-016366febfcd" path="/var/lib/kubelet/pods/a1c5331f-3325-4bab-a0de-016366febfcd/volumes" Mar 17 09:32:18 crc kubenswrapper[4813]: I0317 09:32:18.754481 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d927c167-348f-47f8-9860-b21dac3ca309" path="/var/lib/kubelet/pods/d927c167-348f-47f8-9860-b21dac3ca309/volumes" Mar 17 09:32:18 crc kubenswrapper[4813]: I0317 09:32:18.914738 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92f6c71b-18e8-4015-a231-a741629a4711","Type":"ContainerStarted","Data":"4ccef2d8b50e05342b845d860d7b1a39b82ccca0fd483c11ef1958029df47cfd"} Mar 17 09:32:18 crc kubenswrapper[4813]: I0317 09:32:18.915046 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92f6c71b-18e8-4015-a231-a741629a4711","Type":"ContainerStarted","Data":"d386ba7487283c309d72f0e6c93fd1891a38efa8f8fc1ffec822391cf60f2d28"} Mar 17 09:32:18 crc kubenswrapper[4813]: I0317 09:32:18.917194 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f4b4ddd1-d517-43cc-ac5e-636026efff5d","Type":"ContainerStarted","Data":"953bc4446c49a9e2e9f6cd3ae390bb0fc1d9f521d1c56ca0caf14badb3b224bb"} Mar 17 09:32:18 crc kubenswrapper[4813]: I0317 09:32:18.917228 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f4b4ddd1-d517-43cc-ac5e-636026efff5d","Type":"ContainerStarted","Data":"b3184844bed4947d99d54a6d396f850d94c4d8f719f619c1cb685918cea9b715"} Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.414729 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.588003 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.588089 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-config-data\") pod \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.588134 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-internal-tls-certs\") pod \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.588198 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sh2zr\" (UniqueName: \"kubernetes.io/projected/fc74e31f-b87b-4888-b99f-342c2f72f6b7-kube-api-access-sh2zr\") pod \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.588221 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc74e31f-b87b-4888-b99f-342c2f72f6b7-logs\") pod \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.588309 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-combined-ca-bundle\") pod \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.588339 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fc74e31f-b87b-4888-b99f-342c2f72f6b7-httpd-run\") pod \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.588387 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-scripts\") pod \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\" (UID: \"fc74e31f-b87b-4888-b99f-342c2f72f6b7\") " Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.591055 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc74e31f-b87b-4888-b99f-342c2f72f6b7-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "fc74e31f-b87b-4888-b99f-342c2f72f6b7" (UID: "fc74e31f-b87b-4888-b99f-342c2f72f6b7"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.591129 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fc74e31f-b87b-4888-b99f-342c2f72f6b7-logs" (OuterVolumeSpecName: "logs") pod "fc74e31f-b87b-4888-b99f-342c2f72f6b7" (UID: "fc74e31f-b87b-4888-b99f-342c2f72f6b7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.597107 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "glance") pod "fc74e31f-b87b-4888-b99f-342c2f72f6b7" (UID: "fc74e31f-b87b-4888-b99f-342c2f72f6b7"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.597377 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-scripts" (OuterVolumeSpecName: "scripts") pod "fc74e31f-b87b-4888-b99f-342c2f72f6b7" (UID: "fc74e31f-b87b-4888-b99f-342c2f72f6b7"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.599834 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fc74e31f-b87b-4888-b99f-342c2f72f6b7-kube-api-access-sh2zr" (OuterVolumeSpecName: "kube-api-access-sh2zr") pod "fc74e31f-b87b-4888-b99f-342c2f72f6b7" (UID: "fc74e31f-b87b-4888-b99f-342c2f72f6b7"). InnerVolumeSpecName "kube-api-access-sh2zr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.621226 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fc74e31f-b87b-4888-b99f-342c2f72f6b7" (UID: "fc74e31f-b87b-4888-b99f-342c2f72f6b7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.659063 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-config-data" (OuterVolumeSpecName: "config-data") pod "fc74e31f-b87b-4888-b99f-342c2f72f6b7" (UID: "fc74e31f-b87b-4888-b99f-342c2f72f6b7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.659637 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "fc74e31f-b87b-4888-b99f-342c2f72f6b7" (UID: "fc74e31f-b87b-4888-b99f-342c2f72f6b7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.690537 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sh2zr\" (UniqueName: \"kubernetes.io/projected/fc74e31f-b87b-4888-b99f-342c2f72f6b7-kube-api-access-sh2zr\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.690571 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fc74e31f-b87b-4888-b99f-342c2f72f6b7-logs\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.690581 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.690590 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/fc74e31f-b87b-4888-b99f-342c2f72f6b7-httpd-run\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.690609 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.690649 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.690659 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.690668 4813 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fc74e31f-b87b-4888-b99f-342c2f72f6b7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.709590 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.792288 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.931666 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f4b4ddd1-d517-43cc-ac5e-636026efff5d","Type":"ContainerStarted","Data":"4e8e436e884b5e85187618724c88d18fb2289c5fc6329a5cc44ed4413ad3b158"} Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.936142 4813 generic.go:334] "Generic (PLEG): container finished" podID="fc74e31f-b87b-4888-b99f-342c2f72f6b7" containerID="12e6c7f140dd842351032cd69a0fd18874b77ca813d3ea3f119f9553b1b4b862" exitCode=0 Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.936185 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fc74e31f-b87b-4888-b99f-342c2f72f6b7","Type":"ContainerDied","Data":"12e6c7f140dd842351032cd69a0fd18874b77ca813d3ea3f119f9553b1b4b862"} Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.936215 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"fc74e31f-b87b-4888-b99f-342c2f72f6b7","Type":"ContainerDied","Data":"52d6e9df3f7cb678ef46427439a32bd53667534a016fabfd27a7f8359286cdf3"} Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.936241 4813 scope.go:117] "RemoveContainer" containerID="12e6c7f140dd842351032cd69a0fd18874b77ca813d3ea3f119f9553b1b4b862" Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.936204 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.971234 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.971217095 podStartE2EDuration="3.971217095s" podCreationTimestamp="2026-03-17 09:32:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:32:19.961095607 +0000 UTC m=+1362.061899096" watchObservedRunningTime="2026-03-17 09:32:19.971217095 +0000 UTC m=+1362.072020594" Mar 17 09:32:19 crc kubenswrapper[4813]: I0317 09:32:19.991387 4813 scope.go:117] "RemoveContainer" containerID="20afc3b693977dd502ec4c0428f5b9fd7987ff5309f74555d32af8f5c64e9012" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.014671 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.029677 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.033808 4813 scope.go:117] "RemoveContainer" containerID="12e6c7f140dd842351032cd69a0fd18874b77ca813d3ea3f119f9553b1b4b862" Mar 17 09:32:20 crc kubenswrapper[4813]: E0317 09:32:20.036973 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12e6c7f140dd842351032cd69a0fd18874b77ca813d3ea3f119f9553b1b4b862\": container with ID starting with 12e6c7f140dd842351032cd69a0fd18874b77ca813d3ea3f119f9553b1b4b862 not found: ID does not exist" containerID="12e6c7f140dd842351032cd69a0fd18874b77ca813d3ea3f119f9553b1b4b862" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.037012 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12e6c7f140dd842351032cd69a0fd18874b77ca813d3ea3f119f9553b1b4b862"} err="failed to get container status \"12e6c7f140dd842351032cd69a0fd18874b77ca813d3ea3f119f9553b1b4b862\": rpc error: code = NotFound desc = could not find container \"12e6c7f140dd842351032cd69a0fd18874b77ca813d3ea3f119f9553b1b4b862\": container with ID starting with 12e6c7f140dd842351032cd69a0fd18874b77ca813d3ea3f119f9553b1b4b862 not found: ID does not exist" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.037034 4813 scope.go:117] "RemoveContainer" containerID="20afc3b693977dd502ec4c0428f5b9fd7987ff5309f74555d32af8f5c64e9012" Mar 17 09:32:20 crc kubenswrapper[4813]: E0317 09:32:20.037843 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20afc3b693977dd502ec4c0428f5b9fd7987ff5309f74555d32af8f5c64e9012\": container with ID starting with 20afc3b693977dd502ec4c0428f5b9fd7987ff5309f74555d32af8f5c64e9012 not found: ID does not exist" containerID="20afc3b693977dd502ec4c0428f5b9fd7987ff5309f74555d32af8f5c64e9012" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.037915 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20afc3b693977dd502ec4c0428f5b9fd7987ff5309f74555d32af8f5c64e9012"} err="failed to get container status \"20afc3b693977dd502ec4c0428f5b9fd7987ff5309f74555d32af8f5c64e9012\": rpc error: code = NotFound desc = could not find container \"20afc3b693977dd502ec4c0428f5b9fd7987ff5309f74555d32af8f5c64e9012\": container with ID starting with 20afc3b693977dd502ec4c0428f5b9fd7987ff5309f74555d32af8f5c64e9012 not found: ID does not exist" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.039530 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 09:32:20 crc kubenswrapper[4813]: E0317 09:32:20.040053 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc74e31f-b87b-4888-b99f-342c2f72f6b7" containerName="glance-httpd" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.040074 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc74e31f-b87b-4888-b99f-342c2f72f6b7" containerName="glance-httpd" Mar 17 09:32:20 crc kubenswrapper[4813]: E0317 09:32:20.040103 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fc74e31f-b87b-4888-b99f-342c2f72f6b7" containerName="glance-log" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.040112 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="fc74e31f-b87b-4888-b99f-342c2f72f6b7" containerName="glance-log" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.040343 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc74e31f-b87b-4888-b99f-342c2f72f6b7" containerName="glance-log" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.040364 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="fc74e31f-b87b-4888-b99f-342c2f72f6b7" containerName="glance-httpd" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.042193 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.044299 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.044488 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.047500 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.201859 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7056da71-3a67-4258-8f12-9ab7b50a83ea-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.201900 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tb75\" (UniqueName: \"kubernetes.io/projected/7056da71-3a67-4258-8f12-9ab7b50a83ea-kube-api-access-2tb75\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.201933 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7056da71-3a67-4258-8f12-9ab7b50a83ea-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.201983 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7056da71-3a67-4258-8f12-9ab7b50a83ea-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.202027 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.202057 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7056da71-3a67-4258-8f12-9ab7b50a83ea-logs\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.202073 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7056da71-3a67-4258-8f12-9ab7b50a83ea-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.202093 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7056da71-3a67-4258-8f12-9ab7b50a83ea-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.303274 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7056da71-3a67-4258-8f12-9ab7b50a83ea-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.303336 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.303372 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7056da71-3a67-4258-8f12-9ab7b50a83ea-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.303387 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7056da71-3a67-4258-8f12-9ab7b50a83ea-logs\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.303405 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7056da71-3a67-4258-8f12-9ab7b50a83ea-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.303490 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tb75\" (UniqueName: \"kubernetes.io/projected/7056da71-3a67-4258-8f12-9ab7b50a83ea-kube-api-access-2tb75\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.303508 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7056da71-3a67-4258-8f12-9ab7b50a83ea-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.303536 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7056da71-3a67-4258-8f12-9ab7b50a83ea-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.304437 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7056da71-3a67-4258-8f12-9ab7b50a83ea-logs\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.304514 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.307576 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7056da71-3a67-4258-8f12-9ab7b50a83ea-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.308141 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-85fc85fdfb-n469t" podUID="0a52b24b-cc19-4720-a05b-961015362769" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.156:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.156:8443: connect: connection refused" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.308307 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.308481 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/7056da71-3a67-4258-8f12-9ab7b50a83ea-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.309484 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7056da71-3a67-4258-8f12-9ab7b50a83ea-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.309538 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7056da71-3a67-4258-8f12-9ab7b50a83ea-config-data\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.310010 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7056da71-3a67-4258-8f12-9ab7b50a83ea-scripts\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.329995 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tb75\" (UniqueName: \"kubernetes.io/projected/7056da71-3a67-4258-8f12-9ab7b50a83ea-kube-api-access-2tb75\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.339160 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"glance-default-internal-api-0\" (UID: \"7056da71-3a67-4258-8f12-9ab7b50a83ea\") " pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.363766 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.659299 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-pvfpm"] Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.660595 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pvfpm" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.679404 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-pvfpm"] Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.741199 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fc74e31f-b87b-4888-b99f-342c2f72f6b7" path="/var/lib/kubelet/pods/fc74e31f-b87b-4888-b99f-342c2f72f6b7/volumes" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.753742 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-n2hnd"] Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.761326 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-n2hnd" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.820153 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-8e34-account-create-update-s5l7t"] Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.822159 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xkw4f\" (UniqueName: \"kubernetes.io/projected/8b625cc1-7d38-4a27-b72c-f4ea768de618-kube-api-access-xkw4f\") pod \"nova-api-db-create-pvfpm\" (UID: \"8b625cc1-7d38-4a27-b72c-f4ea768de618\") " pod="openstack/nova-api-db-create-pvfpm" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.822311 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b625cc1-7d38-4a27-b72c-f4ea768de618-operator-scripts\") pod \"nova-api-db-create-pvfpm\" (UID: \"8b625cc1-7d38-4a27-b72c-f4ea768de618\") " pod="openstack/nova-api-db-create-pvfpm" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.830779 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8e34-account-create-update-s5l7t" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.841465 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.912655 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-n2hnd"] Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.924786 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b625cc1-7d38-4a27-b72c-f4ea768de618-operator-scripts\") pod \"nova-api-db-create-pvfpm\" (UID: \"8b625cc1-7d38-4a27-b72c-f4ea768de618\") " pod="openstack/nova-api-db-create-pvfpm" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.924865 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a06cbdf7-9def-426d-8057-35df25d255f6-operator-scripts\") pod \"nova-api-8e34-account-create-update-s5l7t\" (UID: \"a06cbdf7-9def-426d-8057-35df25d255f6\") " pod="openstack/nova-api-8e34-account-create-update-s5l7t" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.924892 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs2tz\" (UniqueName: \"kubernetes.io/projected/7dfd5c4e-760a-47a9-a374-f457c63bc6fa-kube-api-access-xs2tz\") pod \"nova-cell0-db-create-n2hnd\" (UID: \"7dfd5c4e-760a-47a9-a374-f457c63bc6fa\") " pod="openstack/nova-cell0-db-create-n2hnd" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.924976 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sq726\" (UniqueName: \"kubernetes.io/projected/a06cbdf7-9def-426d-8057-35df25d255f6-kube-api-access-sq726\") pod \"nova-api-8e34-account-create-update-s5l7t\" (UID: \"a06cbdf7-9def-426d-8057-35df25d255f6\") " pod="openstack/nova-api-8e34-account-create-update-s5l7t" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.925039 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xkw4f\" (UniqueName: \"kubernetes.io/projected/8b625cc1-7d38-4a27-b72c-f4ea768de618-kube-api-access-xkw4f\") pod \"nova-api-db-create-pvfpm\" (UID: \"8b625cc1-7d38-4a27-b72c-f4ea768de618\") " pod="openstack/nova-api-db-create-pvfpm" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.925057 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7dfd5c4e-760a-47a9-a374-f457c63bc6fa-operator-scripts\") pod \"nova-cell0-db-create-n2hnd\" (UID: \"7dfd5c4e-760a-47a9-a374-f457c63bc6fa\") " pod="openstack/nova-cell0-db-create-n2hnd" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.933509 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b625cc1-7d38-4a27-b72c-f4ea768de618-operator-scripts\") pod \"nova-api-db-create-pvfpm\" (UID: \"8b625cc1-7d38-4a27-b72c-f4ea768de618\") " pod="openstack/nova-api-db-create-pvfpm" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.935735 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8e34-account-create-update-s5l7t"] Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.960347 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xkw4f\" (UniqueName: \"kubernetes.io/projected/8b625cc1-7d38-4a27-b72c-f4ea768de618-kube-api-access-xkw4f\") pod \"nova-api-db-create-pvfpm\" (UID: \"8b625cc1-7d38-4a27-b72c-f4ea768de618\") " pod="openstack/nova-api-db-create-pvfpm" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.975531 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pvfpm" Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.982733 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-vwt8q"] Mar 17 09:32:20 crc kubenswrapper[4813]: I0317 09:32:20.983969 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-vwt8q" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.010690 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92f6c71b-18e8-4015-a231-a741629a4711","Type":"ContainerStarted","Data":"2eb3949ddb871cc652a4e28ce50e6e40979688ec8a5a9d97a0ac2518754b7e3b"} Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.026773 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a06cbdf7-9def-426d-8057-35df25d255f6-operator-scripts\") pod \"nova-api-8e34-account-create-update-s5l7t\" (UID: \"a06cbdf7-9def-426d-8057-35df25d255f6\") " pod="openstack/nova-api-8e34-account-create-update-s5l7t" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.027028 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs2tz\" (UniqueName: \"kubernetes.io/projected/7dfd5c4e-760a-47a9-a374-f457c63bc6fa-kube-api-access-xs2tz\") pod \"nova-cell0-db-create-n2hnd\" (UID: \"7dfd5c4e-760a-47a9-a374-f457c63bc6fa\") " pod="openstack/nova-cell0-db-create-n2hnd" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.027220 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sq726\" (UniqueName: \"kubernetes.io/projected/a06cbdf7-9def-426d-8057-35df25d255f6-kube-api-access-sq726\") pod \"nova-api-8e34-account-create-update-s5l7t\" (UID: \"a06cbdf7-9def-426d-8057-35df25d255f6\") " pod="openstack/nova-api-8e34-account-create-update-s5l7t" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.027390 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7dfd5c4e-760a-47a9-a374-f457c63bc6fa-operator-scripts\") pod \"nova-cell0-db-create-n2hnd\" (UID: \"7dfd5c4e-760a-47a9-a374-f457c63bc6fa\") " pod="openstack/nova-cell0-db-create-n2hnd" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.028288 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7dfd5c4e-760a-47a9-a374-f457c63bc6fa-operator-scripts\") pod \"nova-cell0-db-create-n2hnd\" (UID: \"7dfd5c4e-760a-47a9-a374-f457c63bc6fa\") " pod="openstack/nova-cell0-db-create-n2hnd" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.035528 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-vwt8q"] Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.039037 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a06cbdf7-9def-426d-8057-35df25d255f6-operator-scripts\") pod \"nova-api-8e34-account-create-update-s5l7t\" (UID: \"a06cbdf7-9def-426d-8057-35df25d255f6\") " pod="openstack/nova-api-8e34-account-create-update-s5l7t" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.051372 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs2tz\" (UniqueName: \"kubernetes.io/projected/7dfd5c4e-760a-47a9-a374-f457c63bc6fa-kube-api-access-xs2tz\") pod \"nova-cell0-db-create-n2hnd\" (UID: \"7dfd5c4e-760a-47a9-a374-f457c63bc6fa\") " pod="openstack/nova-cell0-db-create-n2hnd" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.051428 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-f71a-account-create-update-mbsjs"] Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.053039 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f71a-account-create-update-mbsjs" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.054558 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.059470 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sq726\" (UniqueName: \"kubernetes.io/projected/a06cbdf7-9def-426d-8057-35df25d255f6-kube-api-access-sq726\") pod \"nova-api-8e34-account-create-update-s5l7t\" (UID: \"a06cbdf7-9def-426d-8057-35df25d255f6\") " pod="openstack/nova-api-8e34-account-create-update-s5l7t" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.074590 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f71a-account-create-update-mbsjs"] Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.091745 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.120945 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-n2hnd" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.133307 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-69lsz\" (UniqueName: \"kubernetes.io/projected/9534b041-bb7d-48c2-a5e1-ac9ca0af187c-kube-api-access-69lsz\") pod \"nova-cell1-db-create-vwt8q\" (UID: \"9534b041-bb7d-48c2-a5e1-ac9ca0af187c\") " pod="openstack/nova-cell1-db-create-vwt8q" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.133714 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9534b041-bb7d-48c2-a5e1-ac9ca0af187c-operator-scripts\") pod \"nova-cell1-db-create-vwt8q\" (UID: \"9534b041-bb7d-48c2-a5e1-ac9ca0af187c\") " pod="openstack/nova-cell1-db-create-vwt8q" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.163649 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-f51b-account-create-update-sqrd9"] Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.165019 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f51b-account-create-update-sqrd9" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.168452 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.193297 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f51b-account-create-update-sqrd9"] Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.237678 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9534b041-bb7d-48c2-a5e1-ac9ca0af187c-operator-scripts\") pod \"nova-cell1-db-create-vwt8q\" (UID: \"9534b041-bb7d-48c2-a5e1-ac9ca0af187c\") " pod="openstack/nova-cell1-db-create-vwt8q" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.238094 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d67687bd-2db1-4a9d-8869-0aa9c453f5db-operator-scripts\") pod \"nova-cell0-f71a-account-create-update-mbsjs\" (UID: \"d67687bd-2db1-4a9d-8869-0aa9c453f5db\") " pod="openstack/nova-cell0-f71a-account-create-update-mbsjs" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.238138 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-69lsz\" (UniqueName: \"kubernetes.io/projected/9534b041-bb7d-48c2-a5e1-ac9ca0af187c-kube-api-access-69lsz\") pod \"nova-cell1-db-create-vwt8q\" (UID: \"9534b041-bb7d-48c2-a5e1-ac9ca0af187c\") " pod="openstack/nova-cell1-db-create-vwt8q" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.238211 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtzxh\" (UniqueName: \"kubernetes.io/projected/d67687bd-2db1-4a9d-8869-0aa9c453f5db-kube-api-access-xtzxh\") pod \"nova-cell0-f71a-account-create-update-mbsjs\" (UID: \"d67687bd-2db1-4a9d-8869-0aa9c453f5db\") " pod="openstack/nova-cell0-f71a-account-create-update-mbsjs" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.240849 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9534b041-bb7d-48c2-a5e1-ac9ca0af187c-operator-scripts\") pod \"nova-cell1-db-create-vwt8q\" (UID: \"9534b041-bb7d-48c2-a5e1-ac9ca0af187c\") " pod="openstack/nova-cell1-db-create-vwt8q" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.255999 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8e34-account-create-update-s5l7t" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.258121 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-69lsz\" (UniqueName: \"kubernetes.io/projected/9534b041-bb7d-48c2-a5e1-ac9ca0af187c-kube-api-access-69lsz\") pod \"nova-cell1-db-create-vwt8q\" (UID: \"9534b041-bb7d-48c2-a5e1-ac9ca0af187c\") " pod="openstack/nova-cell1-db-create-vwt8q" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.339586 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d67687bd-2db1-4a9d-8869-0aa9c453f5db-operator-scripts\") pod \"nova-cell0-f71a-account-create-update-mbsjs\" (UID: \"d67687bd-2db1-4a9d-8869-0aa9c453f5db\") " pod="openstack/nova-cell0-f71a-account-create-update-mbsjs" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.339645 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56ae9e04-8d01-4211-a2d0-2fc8dfff9593-operator-scripts\") pod \"nova-cell1-f51b-account-create-update-sqrd9\" (UID: \"56ae9e04-8d01-4211-a2d0-2fc8dfff9593\") " pod="openstack/nova-cell1-f51b-account-create-update-sqrd9" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.339687 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtzxh\" (UniqueName: \"kubernetes.io/projected/d67687bd-2db1-4a9d-8869-0aa9c453f5db-kube-api-access-xtzxh\") pod \"nova-cell0-f71a-account-create-update-mbsjs\" (UID: \"d67687bd-2db1-4a9d-8869-0aa9c453f5db\") " pod="openstack/nova-cell0-f71a-account-create-update-mbsjs" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.340387 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8dn8m\" (UniqueName: \"kubernetes.io/projected/56ae9e04-8d01-4211-a2d0-2fc8dfff9593-kube-api-access-8dn8m\") pod \"nova-cell1-f51b-account-create-update-sqrd9\" (UID: \"56ae9e04-8d01-4211-a2d0-2fc8dfff9593\") " pod="openstack/nova-cell1-f51b-account-create-update-sqrd9" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.341681 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d67687bd-2db1-4a9d-8869-0aa9c453f5db-operator-scripts\") pod \"nova-cell0-f71a-account-create-update-mbsjs\" (UID: \"d67687bd-2db1-4a9d-8869-0aa9c453f5db\") " pod="openstack/nova-cell0-f71a-account-create-update-mbsjs" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.357747 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtzxh\" (UniqueName: \"kubernetes.io/projected/d67687bd-2db1-4a9d-8869-0aa9c453f5db-kube-api-access-xtzxh\") pod \"nova-cell0-f71a-account-create-update-mbsjs\" (UID: \"d67687bd-2db1-4a9d-8869-0aa9c453f5db\") " pod="openstack/nova-cell0-f71a-account-create-update-mbsjs" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.384019 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-vwt8q" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.409520 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f71a-account-create-update-mbsjs" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.443705 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56ae9e04-8d01-4211-a2d0-2fc8dfff9593-operator-scripts\") pod \"nova-cell1-f51b-account-create-update-sqrd9\" (UID: \"56ae9e04-8d01-4211-a2d0-2fc8dfff9593\") " pod="openstack/nova-cell1-f51b-account-create-update-sqrd9" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.443844 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8dn8m\" (UniqueName: \"kubernetes.io/projected/56ae9e04-8d01-4211-a2d0-2fc8dfff9593-kube-api-access-8dn8m\") pod \"nova-cell1-f51b-account-create-update-sqrd9\" (UID: \"56ae9e04-8d01-4211-a2d0-2fc8dfff9593\") " pod="openstack/nova-cell1-f51b-account-create-update-sqrd9" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.444797 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56ae9e04-8d01-4211-a2d0-2fc8dfff9593-operator-scripts\") pod \"nova-cell1-f51b-account-create-update-sqrd9\" (UID: \"56ae9e04-8d01-4211-a2d0-2fc8dfff9593\") " pod="openstack/nova-cell1-f51b-account-create-update-sqrd9" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.459859 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8dn8m\" (UniqueName: \"kubernetes.io/projected/56ae9e04-8d01-4211-a2d0-2fc8dfff9593-kube-api-access-8dn8m\") pod \"nova-cell1-f51b-account-create-update-sqrd9\" (UID: \"56ae9e04-8d01-4211-a2d0-2fc8dfff9593\") " pod="openstack/nova-cell1-f51b-account-create-update-sqrd9" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.487106 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f51b-account-create-update-sqrd9" Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.569380 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-pvfpm"] Mar 17 09:32:21 crc kubenswrapper[4813]: W0317 09:32:21.640182 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8b625cc1_7d38_4a27_b72c_f4ea768de618.slice/crio-c9f9b03cede11c6815b98e45341f86e1a276f356c31861cac878e344acf439b6 WatchSource:0}: Error finding container c9f9b03cede11c6815b98e45341f86e1a276f356c31861cac878e344acf439b6: Status 404 returned error can't find the container with id c9f9b03cede11c6815b98e45341f86e1a276f356c31861cac878e344acf439b6 Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.646325 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-8e34-account-create-update-s5l7t"] Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.710443 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-n2hnd"] Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.734731 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-76bd9cc6f7-x2p6w" Mar 17 09:32:21 crc kubenswrapper[4813]: W0317 09:32:21.752054 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7dfd5c4e_760a_47a9_a374_f457c63bc6fa.slice/crio-19aa79e5c7a9eaa64f9a4f961b934645c13f2f5a1b505149501d458535f63567 WatchSource:0}: Error finding container 19aa79e5c7a9eaa64f9a4f961b934645c13f2f5a1b505149501d458535f63567: Status 404 returned error can't find the container with id 19aa79e5c7a9eaa64f9a4f961b934645c13f2f5a1b505149501d458535f63567 Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.813339 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b5495b5bd-g77kp"] Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.813554 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-b5495b5bd-g77kp" podUID="da12db2f-742c-4bf8-8c9e-3ee0868d0c63" containerName="neutron-api" containerID="cri-o://8c5f1c181c9c83af6c9eac398fcc89f4f7500de195959ec00e013725d482ce66" gracePeriod=30 Mar 17 09:32:21 crc kubenswrapper[4813]: I0317 09:32:21.814716 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-b5495b5bd-g77kp" podUID="da12db2f-742c-4bf8-8c9e-3ee0868d0c63" containerName="neutron-httpd" containerID="cri-o://c84fbe3bc0c39f7fff03f9ed586e1ddcb0e72ef85170e94045a5c1c7eabcb226" gracePeriod=30 Mar 17 09:32:22 crc kubenswrapper[4813]: I0317 09:32:22.048155 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92f6c71b-18e8-4015-a231-a741629a4711","Type":"ContainerStarted","Data":"e39df0ae8b176e5ec9935f93b5a9c0f1d1bcf2895a38dfca5b4e67f9beea5969"} Mar 17 09:32:22 crc kubenswrapper[4813]: I0317 09:32:22.050891 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-n2hnd" event={"ID":"7dfd5c4e-760a-47a9-a374-f457c63bc6fa","Type":"ContainerStarted","Data":"acf4ed9b63abf3ea698cbc23be53c44689ccf7056e109aef3577d28d27e86efb"} Mar 17 09:32:22 crc kubenswrapper[4813]: I0317 09:32:22.050940 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-n2hnd" event={"ID":"7dfd5c4e-760a-47a9-a374-f457c63bc6fa","Type":"ContainerStarted","Data":"19aa79e5c7a9eaa64f9a4f961b934645c13f2f5a1b505149501d458535f63567"} Mar 17 09:32:22 crc kubenswrapper[4813]: I0317 09:32:22.054382 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8e34-account-create-update-s5l7t" event={"ID":"a06cbdf7-9def-426d-8057-35df25d255f6","Type":"ContainerStarted","Data":"3103019bec4211362f28316017579ec261a49e21c63ac9db5f43fe82b5be0438"} Mar 17 09:32:22 crc kubenswrapper[4813]: I0317 09:32:22.054411 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8e34-account-create-update-s5l7t" event={"ID":"a06cbdf7-9def-426d-8057-35df25d255f6","Type":"ContainerStarted","Data":"f35c13f168bff0b2ee4fb2366fd45a894d746088d9afae74da770a70a199de4f"} Mar 17 09:32:22 crc kubenswrapper[4813]: I0317 09:32:22.056541 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pvfpm" event={"ID":"8b625cc1-7d38-4a27-b72c-f4ea768de618","Type":"ContainerStarted","Data":"db0de2d2ef92ea95a2c95541079ed04a0eca3b0263df99907a6b1b18d39d3381"} Mar 17 09:32:22 crc kubenswrapper[4813]: I0317 09:32:22.056568 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pvfpm" event={"ID":"8b625cc1-7d38-4a27-b72c-f4ea768de618","Type":"ContainerStarted","Data":"c9f9b03cede11c6815b98e45341f86e1a276f356c31861cac878e344acf439b6"} Mar 17 09:32:22 crc kubenswrapper[4813]: I0317 09:32:22.058307 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7056da71-3a67-4258-8f12-9ab7b50a83ea","Type":"ContainerStarted","Data":"a01a4e277c4e06ab8dce72eb17c0312e6297617c153c892304f6aec39ce6b7c9"} Mar 17 09:32:22 crc kubenswrapper[4813]: I0317 09:32:22.058337 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7056da71-3a67-4258-8f12-9ab7b50a83ea","Type":"ContainerStarted","Data":"17ba9338b095bd6244c1ad20df3121fd90049ee635bf2d7bd19477e180a035e2"} Mar 17 09:32:22 crc kubenswrapper[4813]: I0317 09:32:22.104923 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-n2hnd" podStartSLOduration=2.104904321 podStartE2EDuration="2.104904321s" podCreationTimestamp="2026-03-17 09:32:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:32:22.095117734 +0000 UTC m=+1364.195921233" watchObservedRunningTime="2026-03-17 09:32:22.104904321 +0000 UTC m=+1364.205707820" Mar 17 09:32:22 crc kubenswrapper[4813]: I0317 09:32:22.125578 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-vwt8q"] Mar 17 09:32:22 crc kubenswrapper[4813]: I0317 09:32:22.125954 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-pvfpm" podStartSLOduration=2.125933843 podStartE2EDuration="2.125933843s" podCreationTimestamp="2026-03-17 09:32:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:32:22.119836291 +0000 UTC m=+1364.220639790" watchObservedRunningTime="2026-03-17 09:32:22.125933843 +0000 UTC m=+1364.226737342" Mar 17 09:32:22 crc kubenswrapper[4813]: W0317 09:32:22.166164 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod56ae9e04_8d01_4211_a2d0_2fc8dfff9593.slice/crio-e5ee8d546c2dcb5e3bb114e780939156a5b6e91c9e56e18f51a0e54a12e92b82 WatchSource:0}: Error finding container e5ee8d546c2dcb5e3bb114e780939156a5b6e91c9e56e18f51a0e54a12e92b82: Status 404 returned error can't find the container with id e5ee8d546c2dcb5e3bb114e780939156a5b6e91c9e56e18f51a0e54a12e92b82 Mar 17 09:32:22 crc kubenswrapper[4813]: I0317 09:32:22.166293 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-f51b-account-create-update-sqrd9"] Mar 17 09:32:22 crc kubenswrapper[4813]: I0317 09:32:22.217271 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-f71a-account-create-update-mbsjs"] Mar 17 09:32:22 crc kubenswrapper[4813]: I0317 09:32:22.221501 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-8e34-account-create-update-s5l7t" podStartSLOduration=2.221483737 podStartE2EDuration="2.221483737s" podCreationTimestamp="2026-03-17 09:32:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:32:22.144013421 +0000 UTC m=+1364.244816920" watchObservedRunningTime="2026-03-17 09:32:22.221483737 +0000 UTC m=+1364.322287236" Mar 17 09:32:23 crc kubenswrapper[4813]: I0317 09:32:23.070045 4813 generic.go:334] "Generic (PLEG): container finished" podID="56ae9e04-8d01-4211-a2d0-2fc8dfff9593" containerID="6be42a9854bd3c683909ea09eef0afa2b97a03c03b43cf65dade9d0f6b01af98" exitCode=0 Mar 17 09:32:23 crc kubenswrapper[4813]: I0317 09:32:23.070153 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f51b-account-create-update-sqrd9" event={"ID":"56ae9e04-8d01-4211-a2d0-2fc8dfff9593","Type":"ContainerDied","Data":"6be42a9854bd3c683909ea09eef0afa2b97a03c03b43cf65dade9d0f6b01af98"} Mar 17 09:32:23 crc kubenswrapper[4813]: I0317 09:32:23.070666 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f51b-account-create-update-sqrd9" event={"ID":"56ae9e04-8d01-4211-a2d0-2fc8dfff9593","Type":"ContainerStarted","Data":"e5ee8d546c2dcb5e3bb114e780939156a5b6e91c9e56e18f51a0e54a12e92b82"} Mar 17 09:32:23 crc kubenswrapper[4813]: I0317 09:32:23.072861 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"7056da71-3a67-4258-8f12-9ab7b50a83ea","Type":"ContainerStarted","Data":"fb2a12372d8075d6bcc40c3f52fd3e8ab69c5325c3c5cfc8a6b6709198752884"} Mar 17 09:32:23 crc kubenswrapper[4813]: I0317 09:32:23.074804 4813 generic.go:334] "Generic (PLEG): container finished" podID="7dfd5c4e-760a-47a9-a374-f457c63bc6fa" containerID="acf4ed9b63abf3ea698cbc23be53c44689ccf7056e109aef3577d28d27e86efb" exitCode=0 Mar 17 09:32:23 crc kubenswrapper[4813]: I0317 09:32:23.074835 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-n2hnd" event={"ID":"7dfd5c4e-760a-47a9-a374-f457c63bc6fa","Type":"ContainerDied","Data":"acf4ed9b63abf3ea698cbc23be53c44689ccf7056e109aef3577d28d27e86efb"} Mar 17 09:32:23 crc kubenswrapper[4813]: I0317 09:32:23.076760 4813 generic.go:334] "Generic (PLEG): container finished" podID="d67687bd-2db1-4a9d-8869-0aa9c453f5db" containerID="10d6ff5ffda0d5731232121b139d9ec7456c86e1ff460c20c0860d546c309749" exitCode=0 Mar 17 09:32:23 crc kubenswrapper[4813]: I0317 09:32:23.076830 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f71a-account-create-update-mbsjs" event={"ID":"d67687bd-2db1-4a9d-8869-0aa9c453f5db","Type":"ContainerDied","Data":"10d6ff5ffda0d5731232121b139d9ec7456c86e1ff460c20c0860d546c309749"} Mar 17 09:32:23 crc kubenswrapper[4813]: I0317 09:32:23.076857 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f71a-account-create-update-mbsjs" event={"ID":"d67687bd-2db1-4a9d-8869-0aa9c453f5db","Type":"ContainerStarted","Data":"cbd0d4e92d32860a97c2181cda902e243b09db999e8b2b75e7bf16c0ef5ff9ec"} Mar 17 09:32:23 crc kubenswrapper[4813]: I0317 09:32:23.078345 4813 generic.go:334] "Generic (PLEG): container finished" podID="9534b041-bb7d-48c2-a5e1-ac9ca0af187c" containerID="e9fd1c7b79e5b326c1264811d9cb8729ac74fed9395c7bced7ae92947a232cdb" exitCode=0 Mar 17 09:32:23 crc kubenswrapper[4813]: I0317 09:32:23.078388 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-vwt8q" event={"ID":"9534b041-bb7d-48c2-a5e1-ac9ca0af187c","Type":"ContainerDied","Data":"e9fd1c7b79e5b326c1264811d9cb8729ac74fed9395c7bced7ae92947a232cdb"} Mar 17 09:32:23 crc kubenswrapper[4813]: I0317 09:32:23.078433 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-vwt8q" event={"ID":"9534b041-bb7d-48c2-a5e1-ac9ca0af187c","Type":"ContainerStarted","Data":"a4b1ff9d25ac9423700e268cec967bc8e4f8e29dde171894a36058651c5f8898"} Mar 17 09:32:23 crc kubenswrapper[4813]: I0317 09:32:23.080480 4813 generic.go:334] "Generic (PLEG): container finished" podID="a06cbdf7-9def-426d-8057-35df25d255f6" containerID="3103019bec4211362f28316017579ec261a49e21c63ac9db5f43fe82b5be0438" exitCode=0 Mar 17 09:32:23 crc kubenswrapper[4813]: I0317 09:32:23.080539 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8e34-account-create-update-s5l7t" event={"ID":"a06cbdf7-9def-426d-8057-35df25d255f6","Type":"ContainerDied","Data":"3103019bec4211362f28316017579ec261a49e21c63ac9db5f43fe82b5be0438"} Mar 17 09:32:23 crc kubenswrapper[4813]: I0317 09:32:23.082348 4813 generic.go:334] "Generic (PLEG): container finished" podID="8b625cc1-7d38-4a27-b72c-f4ea768de618" containerID="db0de2d2ef92ea95a2c95541079ed04a0eca3b0263df99907a6b1b18d39d3381" exitCode=0 Mar 17 09:32:23 crc kubenswrapper[4813]: I0317 09:32:23.082395 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pvfpm" event={"ID":"8b625cc1-7d38-4a27-b72c-f4ea768de618","Type":"ContainerDied","Data":"db0de2d2ef92ea95a2c95541079ed04a0eca3b0263df99907a6b1b18d39d3381"} Mar 17 09:32:23 crc kubenswrapper[4813]: I0317 09:32:23.084578 4813 generic.go:334] "Generic (PLEG): container finished" podID="da12db2f-742c-4bf8-8c9e-3ee0868d0c63" containerID="c84fbe3bc0c39f7fff03f9ed586e1ddcb0e72ef85170e94045a5c1c7eabcb226" exitCode=0 Mar 17 09:32:23 crc kubenswrapper[4813]: I0317 09:32:23.084625 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b5495b5bd-g77kp" event={"ID":"da12db2f-742c-4bf8-8c9e-3ee0868d0c63","Type":"ContainerDied","Data":"c84fbe3bc0c39f7fff03f9ed586e1ddcb0e72ef85170e94045a5c1c7eabcb226"} Mar 17 09:32:23 crc kubenswrapper[4813]: I0317 09:32:23.169869 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.169845455 podStartE2EDuration="3.169845455s" podCreationTimestamp="2026-03-17 09:32:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:32:23.167029856 +0000 UTC m=+1365.267833355" watchObservedRunningTime="2026-03-17 09:32:23.169845455 +0000 UTC m=+1365.270648954" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.098835 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92f6c71b-18e8-4015-a231-a741629a4711","Type":"ContainerStarted","Data":"44163684427cfcdba1b739c304593eb2d1a04589d09d1b0e2ca1127abcd26d4b"} Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.102213 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92f6c71b-18e8-4015-a231-a741629a4711" containerName="proxy-httpd" containerID="cri-o://44163684427cfcdba1b739c304593eb2d1a04589d09d1b0e2ca1127abcd26d4b" gracePeriod=30 Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.102536 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92f6c71b-18e8-4015-a231-a741629a4711" containerName="sg-core" containerID="cri-o://e39df0ae8b176e5ec9935f93b5a9c0f1d1bcf2895a38dfca5b4e67f9beea5969" gracePeriod=30 Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.102586 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92f6c71b-18e8-4015-a231-a741629a4711" containerName="ceilometer-notification-agent" containerID="cri-o://2eb3949ddb871cc652a4e28ce50e6e40979688ec8a5a9d97a0ac2518754b7e3b" gracePeriod=30 Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.103929 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="92f6c71b-18e8-4015-a231-a741629a4711" containerName="ceilometer-central-agent" containerID="cri-o://4ccef2d8b50e05342b845d860d7b1a39b82ccca0fd483c11ef1958029df47cfd" gracePeriod=30 Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.154182 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.971700439 podStartE2EDuration="8.154163012s" podCreationTimestamp="2026-03-17 09:32:16 +0000 UTC" firstStartedPulling="2026-03-17 09:32:17.980552326 +0000 UTC m=+1360.081355825" lastFinishedPulling="2026-03-17 09:32:23.163014899 +0000 UTC m=+1365.263818398" observedRunningTime="2026-03-17 09:32:24.145564611 +0000 UTC m=+1366.246368110" watchObservedRunningTime="2026-03-17 09:32:24.154163012 +0000 UTC m=+1366.254966511" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.626304 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f71a-account-create-update-mbsjs" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.715584 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtzxh\" (UniqueName: \"kubernetes.io/projected/d67687bd-2db1-4a9d-8869-0aa9c453f5db-kube-api-access-xtzxh\") pod \"d67687bd-2db1-4a9d-8869-0aa9c453f5db\" (UID: \"d67687bd-2db1-4a9d-8869-0aa9c453f5db\") " Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.715673 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d67687bd-2db1-4a9d-8869-0aa9c453f5db-operator-scripts\") pod \"d67687bd-2db1-4a9d-8869-0aa9c453f5db\" (UID: \"d67687bd-2db1-4a9d-8869-0aa9c453f5db\") " Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.716962 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d67687bd-2db1-4a9d-8869-0aa9c453f5db-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d67687bd-2db1-4a9d-8869-0aa9c453f5db" (UID: "d67687bd-2db1-4a9d-8869-0aa9c453f5db"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.724694 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d67687bd-2db1-4a9d-8869-0aa9c453f5db-kube-api-access-xtzxh" (OuterVolumeSpecName: "kube-api-access-xtzxh") pod "d67687bd-2db1-4a9d-8869-0aa9c453f5db" (UID: "d67687bd-2db1-4a9d-8869-0aa9c453f5db"). InnerVolumeSpecName "kube-api-access-xtzxh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.729041 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f51b-account-create-update-sqrd9" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.735561 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-n2hnd" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.748542 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-vwt8q" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.767512 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pvfpm" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.780393 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8e34-account-create-update-s5l7t" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.832277 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7dfd5c4e-760a-47a9-a374-f457c63bc6fa-operator-scripts\") pod \"7dfd5c4e-760a-47a9-a374-f457c63bc6fa\" (UID: \"7dfd5c4e-760a-47a9-a374-f457c63bc6fa\") " Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.832350 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xkw4f\" (UniqueName: \"kubernetes.io/projected/8b625cc1-7d38-4a27-b72c-f4ea768de618-kube-api-access-xkw4f\") pod \"8b625cc1-7d38-4a27-b72c-f4ea768de618\" (UID: \"8b625cc1-7d38-4a27-b72c-f4ea768de618\") " Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.832437 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9534b041-bb7d-48c2-a5e1-ac9ca0af187c-operator-scripts\") pod \"9534b041-bb7d-48c2-a5e1-ac9ca0af187c\" (UID: \"9534b041-bb7d-48c2-a5e1-ac9ca0af187c\") " Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.832487 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sq726\" (UniqueName: \"kubernetes.io/projected/a06cbdf7-9def-426d-8057-35df25d255f6-kube-api-access-sq726\") pod \"a06cbdf7-9def-426d-8057-35df25d255f6\" (UID: \"a06cbdf7-9def-426d-8057-35df25d255f6\") " Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.832546 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b625cc1-7d38-4a27-b72c-f4ea768de618-operator-scripts\") pod \"8b625cc1-7d38-4a27-b72c-f4ea768de618\" (UID: \"8b625cc1-7d38-4a27-b72c-f4ea768de618\") " Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.832580 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xs2tz\" (UniqueName: \"kubernetes.io/projected/7dfd5c4e-760a-47a9-a374-f457c63bc6fa-kube-api-access-xs2tz\") pod \"7dfd5c4e-760a-47a9-a374-f457c63bc6fa\" (UID: \"7dfd5c4e-760a-47a9-a374-f457c63bc6fa\") " Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.832665 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8dn8m\" (UniqueName: \"kubernetes.io/projected/56ae9e04-8d01-4211-a2d0-2fc8dfff9593-kube-api-access-8dn8m\") pod \"56ae9e04-8d01-4211-a2d0-2fc8dfff9593\" (UID: \"56ae9e04-8d01-4211-a2d0-2fc8dfff9593\") " Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.832734 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-69lsz\" (UniqueName: \"kubernetes.io/projected/9534b041-bb7d-48c2-a5e1-ac9ca0af187c-kube-api-access-69lsz\") pod \"9534b041-bb7d-48c2-a5e1-ac9ca0af187c\" (UID: \"9534b041-bb7d-48c2-a5e1-ac9ca0af187c\") " Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.832769 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a06cbdf7-9def-426d-8057-35df25d255f6-operator-scripts\") pod \"a06cbdf7-9def-426d-8057-35df25d255f6\" (UID: \"a06cbdf7-9def-426d-8057-35df25d255f6\") " Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.832791 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56ae9e04-8d01-4211-a2d0-2fc8dfff9593-operator-scripts\") pod \"56ae9e04-8d01-4211-a2d0-2fc8dfff9593\" (UID: \"56ae9e04-8d01-4211-a2d0-2fc8dfff9593\") " Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.832974 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7dfd5c4e-760a-47a9-a374-f457c63bc6fa-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "7dfd5c4e-760a-47a9-a374-f457c63bc6fa" (UID: "7dfd5c4e-760a-47a9-a374-f457c63bc6fa"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.833259 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9534b041-bb7d-48c2-a5e1-ac9ca0af187c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9534b041-bb7d-48c2-a5e1-ac9ca0af187c" (UID: "9534b041-bb7d-48c2-a5e1-ac9ca0af187c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.833498 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtzxh\" (UniqueName: \"kubernetes.io/projected/d67687bd-2db1-4a9d-8869-0aa9c453f5db-kube-api-access-xtzxh\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.833519 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d67687bd-2db1-4a9d-8869-0aa9c453f5db-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.833528 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/7dfd5c4e-760a-47a9-a374-f457c63bc6fa-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.833538 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9534b041-bb7d-48c2-a5e1-ac9ca0af187c-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.834143 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a06cbdf7-9def-426d-8057-35df25d255f6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a06cbdf7-9def-426d-8057-35df25d255f6" (UID: "a06cbdf7-9def-426d-8057-35df25d255f6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.834863 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/56ae9e04-8d01-4211-a2d0-2fc8dfff9593-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "56ae9e04-8d01-4211-a2d0-2fc8dfff9593" (UID: "56ae9e04-8d01-4211-a2d0-2fc8dfff9593"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.834923 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8b625cc1-7d38-4a27-b72c-f4ea768de618-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8b625cc1-7d38-4a27-b72c-f4ea768de618" (UID: "8b625cc1-7d38-4a27-b72c-f4ea768de618"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.840499 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b625cc1-7d38-4a27-b72c-f4ea768de618-kube-api-access-xkw4f" (OuterVolumeSpecName: "kube-api-access-xkw4f") pod "8b625cc1-7d38-4a27-b72c-f4ea768de618" (UID: "8b625cc1-7d38-4a27-b72c-f4ea768de618"). InnerVolumeSpecName "kube-api-access-xkw4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.840634 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a06cbdf7-9def-426d-8057-35df25d255f6-kube-api-access-sq726" (OuterVolumeSpecName: "kube-api-access-sq726") pod "a06cbdf7-9def-426d-8057-35df25d255f6" (UID: "a06cbdf7-9def-426d-8057-35df25d255f6"). InnerVolumeSpecName "kube-api-access-sq726". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.843372 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9534b041-bb7d-48c2-a5e1-ac9ca0af187c-kube-api-access-69lsz" (OuterVolumeSpecName: "kube-api-access-69lsz") pod "9534b041-bb7d-48c2-a5e1-ac9ca0af187c" (UID: "9534b041-bb7d-48c2-a5e1-ac9ca0af187c"). InnerVolumeSpecName "kube-api-access-69lsz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.845682 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dfd5c4e-760a-47a9-a374-f457c63bc6fa-kube-api-access-xs2tz" (OuterVolumeSpecName: "kube-api-access-xs2tz") pod "7dfd5c4e-760a-47a9-a374-f457c63bc6fa" (UID: "7dfd5c4e-760a-47a9-a374-f457c63bc6fa"). InnerVolumeSpecName "kube-api-access-xs2tz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.855261 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56ae9e04-8d01-4211-a2d0-2fc8dfff9593-kube-api-access-8dn8m" (OuterVolumeSpecName: "kube-api-access-8dn8m") pod "56ae9e04-8d01-4211-a2d0-2fc8dfff9593" (UID: "56ae9e04-8d01-4211-a2d0-2fc8dfff9593"). InnerVolumeSpecName "kube-api-access-8dn8m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.934936 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8b625cc1-7d38-4a27-b72c-f4ea768de618-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.934967 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xs2tz\" (UniqueName: \"kubernetes.io/projected/7dfd5c4e-760a-47a9-a374-f457c63bc6fa-kube-api-access-xs2tz\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.934978 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8dn8m\" (UniqueName: \"kubernetes.io/projected/56ae9e04-8d01-4211-a2d0-2fc8dfff9593-kube-api-access-8dn8m\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.934986 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-69lsz\" (UniqueName: \"kubernetes.io/projected/9534b041-bb7d-48c2-a5e1-ac9ca0af187c-kube-api-access-69lsz\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.934995 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a06cbdf7-9def-426d-8057-35df25d255f6-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.935004 4813 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/56ae9e04-8d01-4211-a2d0-2fc8dfff9593-operator-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.935012 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xkw4f\" (UniqueName: \"kubernetes.io/projected/8b625cc1-7d38-4a27-b72c-f4ea768de618-kube-api-access-xkw4f\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:24 crc kubenswrapper[4813]: I0317 09:32:24.935020 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sq726\" (UniqueName: \"kubernetes.io/projected/a06cbdf7-9def-426d-8057-35df25d255f6-kube-api-access-sq726\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.109984 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-n2hnd" event={"ID":"7dfd5c4e-760a-47a9-a374-f457c63bc6fa","Type":"ContainerDied","Data":"19aa79e5c7a9eaa64f9a4f961b934645c13f2f5a1b505149501d458535f63567"} Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.110034 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19aa79e5c7a9eaa64f9a4f961b934645c13f2f5a1b505149501d458535f63567" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.110107 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-n2hnd" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.120852 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-f71a-account-create-update-mbsjs" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.121116 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-f71a-account-create-update-mbsjs" event={"ID":"d67687bd-2db1-4a9d-8869-0aa9c453f5db","Type":"ContainerDied","Data":"cbd0d4e92d32860a97c2181cda902e243b09db999e8b2b75e7bf16c0ef5ff9ec"} Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.121167 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cbd0d4e92d32860a97c2181cda902e243b09db999e8b2b75e7bf16c0ef5ff9ec" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.122880 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-vwt8q" event={"ID":"9534b041-bb7d-48c2-a5e1-ac9ca0af187c","Type":"ContainerDied","Data":"a4b1ff9d25ac9423700e268cec967bc8e4f8e29dde171894a36058651c5f8898"} Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.122917 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a4b1ff9d25ac9423700e268cec967bc8e4f8e29dde171894a36058651c5f8898" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.122893 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-vwt8q" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.124271 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-8e34-account-create-update-s5l7t" event={"ID":"a06cbdf7-9def-426d-8057-35df25d255f6","Type":"ContainerDied","Data":"f35c13f168bff0b2ee4fb2366fd45a894d746088d9afae74da770a70a199de4f"} Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.124302 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f35c13f168bff0b2ee4fb2366fd45a894d746088d9afae74da770a70a199de4f" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.124850 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-8e34-account-create-update-s5l7t" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.126004 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-pvfpm" event={"ID":"8b625cc1-7d38-4a27-b72c-f4ea768de618","Type":"ContainerDied","Data":"c9f9b03cede11c6815b98e45341f86e1a276f356c31861cac878e344acf439b6"} Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.126038 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9f9b03cede11c6815b98e45341f86e1a276f356c31861cac878e344acf439b6" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.126088 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-pvfpm" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.130308 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-f51b-account-create-update-sqrd9" event={"ID":"56ae9e04-8d01-4211-a2d0-2fc8dfff9593","Type":"ContainerDied","Data":"e5ee8d546c2dcb5e3bb114e780939156a5b6e91c9e56e18f51a0e54a12e92b82"} Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.130352 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5ee8d546c2dcb5e3bb114e780939156a5b6e91c9e56e18f51a0e54a12e92b82" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.130408 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-f51b-account-create-update-sqrd9" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.140379 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92f6c71b-18e8-4015-a231-a741629a4711","Type":"ContainerDied","Data":"44163684427cfcdba1b739c304593eb2d1a04589d09d1b0e2ca1127abcd26d4b"} Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.144667 4813 generic.go:334] "Generic (PLEG): container finished" podID="92f6c71b-18e8-4015-a231-a741629a4711" containerID="44163684427cfcdba1b739c304593eb2d1a04589d09d1b0e2ca1127abcd26d4b" exitCode=0 Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.144758 4813 generic.go:334] "Generic (PLEG): container finished" podID="92f6c71b-18e8-4015-a231-a741629a4711" containerID="e39df0ae8b176e5ec9935f93b5a9c0f1d1bcf2895a38dfca5b4e67f9beea5969" exitCode=2 Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.144768 4813 generic.go:334] "Generic (PLEG): container finished" podID="92f6c71b-18e8-4015-a231-a741629a4711" containerID="2eb3949ddb871cc652a4e28ce50e6e40979688ec8a5a9d97a0ac2518754b7e3b" exitCode=0 Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.144799 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92f6c71b-18e8-4015-a231-a741629a4711","Type":"ContainerDied","Data":"e39df0ae8b176e5ec9935f93b5a9c0f1d1bcf2895a38dfca5b4e67f9beea5969"} Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.144837 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92f6c71b-18e8-4015-a231-a741629a4711","Type":"ContainerDied","Data":"2eb3949ddb871cc652a4e28ce50e6e40979688ec8a5a9d97a0ac2518754b7e3b"} Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.675775 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.746314 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0a52b24b-cc19-4720-a05b-961015362769-scripts\") pod \"0a52b24b-cc19-4720-a05b-961015362769\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.746361 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0a52b24b-cc19-4720-a05b-961015362769-horizon-secret-key\") pod \"0a52b24b-cc19-4720-a05b-961015362769\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.746398 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fxwcd\" (UniqueName: \"kubernetes.io/projected/0a52b24b-cc19-4720-a05b-961015362769-kube-api-access-fxwcd\") pod \"0a52b24b-cc19-4720-a05b-961015362769\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.746453 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0a52b24b-cc19-4720-a05b-961015362769-config-data\") pod \"0a52b24b-cc19-4720-a05b-961015362769\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.746505 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a52b24b-cc19-4720-a05b-961015362769-horizon-tls-certs\") pod \"0a52b24b-cc19-4720-a05b-961015362769\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.746549 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a52b24b-cc19-4720-a05b-961015362769-combined-ca-bundle\") pod \"0a52b24b-cc19-4720-a05b-961015362769\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.746586 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a52b24b-cc19-4720-a05b-961015362769-logs\") pod \"0a52b24b-cc19-4720-a05b-961015362769\" (UID: \"0a52b24b-cc19-4720-a05b-961015362769\") " Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.747007 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0a52b24b-cc19-4720-a05b-961015362769-logs" (OuterVolumeSpecName: "logs") pod "0a52b24b-cc19-4720-a05b-961015362769" (UID: "0a52b24b-cc19-4720-a05b-961015362769"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.795045 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a52b24b-cc19-4720-a05b-961015362769-kube-api-access-fxwcd" (OuterVolumeSpecName: "kube-api-access-fxwcd") pod "0a52b24b-cc19-4720-a05b-961015362769" (UID: "0a52b24b-cc19-4720-a05b-961015362769"). InnerVolumeSpecName "kube-api-access-fxwcd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.795813 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a52b24b-cc19-4720-a05b-961015362769-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "0a52b24b-cc19-4720-a05b-961015362769" (UID: "0a52b24b-cc19-4720-a05b-961015362769"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.797565 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a52b24b-cc19-4720-a05b-961015362769-config-data" (OuterVolumeSpecName: "config-data") pod "0a52b24b-cc19-4720-a05b-961015362769" (UID: "0a52b24b-cc19-4720-a05b-961015362769"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.816703 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a52b24b-cc19-4720-a05b-961015362769-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0a52b24b-cc19-4720-a05b-961015362769" (UID: "0a52b24b-cc19-4720-a05b-961015362769"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.821037 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0a52b24b-cc19-4720-a05b-961015362769-scripts" (OuterVolumeSpecName: "scripts") pod "0a52b24b-cc19-4720-a05b-961015362769" (UID: "0a52b24b-cc19-4720-a05b-961015362769"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.828816 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0a52b24b-cc19-4720-a05b-961015362769-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "0a52b24b-cc19-4720-a05b-961015362769" (UID: "0a52b24b-cc19-4720-a05b-961015362769"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.848015 4813 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0a52b24b-cc19-4720-a05b-961015362769-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.848101 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0a52b24b-cc19-4720-a05b-961015362769-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.848116 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0a52b24b-cc19-4720-a05b-961015362769-logs\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.848126 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0a52b24b-cc19-4720-a05b-961015362769-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.848134 4813 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0a52b24b-cc19-4720-a05b-961015362769-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.848142 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fxwcd\" (UniqueName: \"kubernetes.io/projected/0a52b24b-cc19-4720-a05b-961015362769-kube-api-access-fxwcd\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:25 crc kubenswrapper[4813]: I0317 09:32:25.848151 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0a52b24b-cc19-4720-a05b-961015362769-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.160673 4813 generic.go:334] "Generic (PLEG): container finished" podID="0a52b24b-cc19-4720-a05b-961015362769" containerID="f4b4b9a07ddaca0201104a913247a2025ba56488f789d4d1b11072bb55859ee4" exitCode=137 Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.160770 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85fc85fdfb-n469t" event={"ID":"0a52b24b-cc19-4720-a05b-961015362769","Type":"ContainerDied","Data":"f4b4b9a07ddaca0201104a913247a2025ba56488f789d4d1b11072bb55859ee4"} Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.160809 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-85fc85fdfb-n469t" event={"ID":"0a52b24b-cc19-4720-a05b-961015362769","Type":"ContainerDied","Data":"e4ff8b58e4f844a4867d183f3d7529361e570d7af0068b7705e4b87b56b07d24"} Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.160837 4813 scope.go:117] "RemoveContainer" containerID="77e6158b584698861562f3c366b15b1b09acecaf7c7eb2ee5edd3c9ddec18130" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.161009 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-85fc85fdfb-n469t" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.168209 4813 generic.go:334] "Generic (PLEG): container finished" podID="da12db2f-742c-4bf8-8c9e-3ee0868d0c63" containerID="8c5f1c181c9c83af6c9eac398fcc89f4f7500de195959ec00e013725d482ce66" exitCode=0 Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.168404 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b5495b5bd-g77kp" event={"ID":"da12db2f-742c-4bf8-8c9e-3ee0868d0c63","Type":"ContainerDied","Data":"8c5f1c181c9c83af6c9eac398fcc89f4f7500de195959ec00e013725d482ce66"} Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.221978 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-85fc85fdfb-n469t"] Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.233139 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-85fc85fdfb-n469t"] Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.348576 4813 scope.go:117] "RemoveContainer" containerID="f4b4b9a07ddaca0201104a913247a2025ba56488f789d4d1b11072bb55859ee4" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.466726 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jdtxs"] Mar 17 09:32:26 crc kubenswrapper[4813]: E0317 09:32:26.467061 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a52b24b-cc19-4720-a05b-961015362769" containerName="horizon" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.467075 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a52b24b-cc19-4720-a05b-961015362769" containerName="horizon" Mar 17 09:32:26 crc kubenswrapper[4813]: E0317 09:32:26.467098 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a52b24b-cc19-4720-a05b-961015362769" containerName="horizon-log" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.467104 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a52b24b-cc19-4720-a05b-961015362769" containerName="horizon-log" Mar 17 09:32:26 crc kubenswrapper[4813]: E0317 09:32:26.467113 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dfd5c4e-760a-47a9-a374-f457c63bc6fa" containerName="mariadb-database-create" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.467119 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dfd5c4e-760a-47a9-a374-f457c63bc6fa" containerName="mariadb-database-create" Mar 17 09:32:26 crc kubenswrapper[4813]: E0317 09:32:26.467130 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56ae9e04-8d01-4211-a2d0-2fc8dfff9593" containerName="mariadb-account-create-update" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.467136 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="56ae9e04-8d01-4211-a2d0-2fc8dfff9593" containerName="mariadb-account-create-update" Mar 17 09:32:26 crc kubenswrapper[4813]: E0317 09:32:26.467146 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b625cc1-7d38-4a27-b72c-f4ea768de618" containerName="mariadb-database-create" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.467154 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b625cc1-7d38-4a27-b72c-f4ea768de618" containerName="mariadb-database-create" Mar 17 09:32:26 crc kubenswrapper[4813]: E0317 09:32:26.467167 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d67687bd-2db1-4a9d-8869-0aa9c453f5db" containerName="mariadb-account-create-update" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.467172 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d67687bd-2db1-4a9d-8869-0aa9c453f5db" containerName="mariadb-account-create-update" Mar 17 09:32:26 crc kubenswrapper[4813]: E0317 09:32:26.467180 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9534b041-bb7d-48c2-a5e1-ac9ca0af187c" containerName="mariadb-database-create" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.467186 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9534b041-bb7d-48c2-a5e1-ac9ca0af187c" containerName="mariadb-database-create" Mar 17 09:32:26 crc kubenswrapper[4813]: E0317 09:32:26.467199 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a06cbdf7-9def-426d-8057-35df25d255f6" containerName="mariadb-account-create-update" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.467205 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a06cbdf7-9def-426d-8057-35df25d255f6" containerName="mariadb-account-create-update" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.467352 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9534b041-bb7d-48c2-a5e1-ac9ca0af187c" containerName="mariadb-database-create" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.467365 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a52b24b-cc19-4720-a05b-961015362769" containerName="horizon-log" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.467371 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a52b24b-cc19-4720-a05b-961015362769" containerName="horizon" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.467380 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dfd5c4e-760a-47a9-a374-f457c63bc6fa" containerName="mariadb-database-create" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.467388 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d67687bd-2db1-4a9d-8869-0aa9c453f5db" containerName="mariadb-account-create-update" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.467401 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b625cc1-7d38-4a27-b72c-f4ea768de618" containerName="mariadb-database-create" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.467410 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="56ae9e04-8d01-4211-a2d0-2fc8dfff9593" containerName="mariadb-account-create-update" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.467421 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a06cbdf7-9def-426d-8057-35df25d255f6" containerName="mariadb-account-create-update" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.467988 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jdtxs" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.469927 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.470752 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.470818 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-k8fzw" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.488009 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b5495b5bd-g77kp" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.491715 4813 scope.go:117] "RemoveContainer" containerID="77e6158b584698861562f3c366b15b1b09acecaf7c7eb2ee5edd3c9ddec18130" Mar 17 09:32:26 crc kubenswrapper[4813]: E0317 09:32:26.492015 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"77e6158b584698861562f3c366b15b1b09acecaf7c7eb2ee5edd3c9ddec18130\": container with ID starting with 77e6158b584698861562f3c366b15b1b09acecaf7c7eb2ee5edd3c9ddec18130 not found: ID does not exist" containerID="77e6158b584698861562f3c366b15b1b09acecaf7c7eb2ee5edd3c9ddec18130" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.492044 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"77e6158b584698861562f3c366b15b1b09acecaf7c7eb2ee5edd3c9ddec18130"} err="failed to get container status \"77e6158b584698861562f3c366b15b1b09acecaf7c7eb2ee5edd3c9ddec18130\": rpc error: code = NotFound desc = could not find container \"77e6158b584698861562f3c366b15b1b09acecaf7c7eb2ee5edd3c9ddec18130\": container with ID starting with 77e6158b584698861562f3c366b15b1b09acecaf7c7eb2ee5edd3c9ddec18130 not found: ID does not exist" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.492065 4813 scope.go:117] "RemoveContainer" containerID="f4b4b9a07ddaca0201104a913247a2025ba56488f789d4d1b11072bb55859ee4" Mar 17 09:32:26 crc kubenswrapper[4813]: E0317 09:32:26.492295 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4b4b9a07ddaca0201104a913247a2025ba56488f789d4d1b11072bb55859ee4\": container with ID starting with f4b4b9a07ddaca0201104a913247a2025ba56488f789d4d1b11072bb55859ee4 not found: ID does not exist" containerID="f4b4b9a07ddaca0201104a913247a2025ba56488f789d4d1b11072bb55859ee4" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.492320 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4b4b9a07ddaca0201104a913247a2025ba56488f789d4d1b11072bb55859ee4"} err="failed to get container status \"f4b4b9a07ddaca0201104a913247a2025ba56488f789d4d1b11072bb55859ee4\": rpc error: code = NotFound desc = could not find container \"f4b4b9a07ddaca0201104a913247a2025ba56488f789d4d1b11072bb55859ee4\": container with ID starting with f4b4b9a07ddaca0201104a913247a2025ba56488f789d4d1b11072bb55859ee4 not found: ID does not exist" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.509359 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jdtxs"] Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.662059 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-ovndb-tls-certs\") pod \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\" (UID: \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\") " Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.662125 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9fxk\" (UniqueName: \"kubernetes.io/projected/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-kube-api-access-v9fxk\") pod \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\" (UID: \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\") " Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.662166 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-combined-ca-bundle\") pod \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\" (UID: \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\") " Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.662187 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-httpd-config\") pod \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\" (UID: \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\") " Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.662318 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-config\") pod \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\" (UID: \"da12db2f-742c-4bf8-8c9e-3ee0868d0c63\") " Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.662585 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzgvx\" (UniqueName: \"kubernetes.io/projected/48793e32-107b-4ba7-b3a6-3e8172306408-kube-api-access-nzgvx\") pod \"nova-cell0-conductor-db-sync-jdtxs\" (UID: \"48793e32-107b-4ba7-b3a6-3e8172306408\") " pod="openstack/nova-cell0-conductor-db-sync-jdtxs" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.662699 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48793e32-107b-4ba7-b3a6-3e8172306408-scripts\") pod \"nova-cell0-conductor-db-sync-jdtxs\" (UID: \"48793e32-107b-4ba7-b3a6-3e8172306408\") " pod="openstack/nova-cell0-conductor-db-sync-jdtxs" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.662752 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48793e32-107b-4ba7-b3a6-3e8172306408-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jdtxs\" (UID: \"48793e32-107b-4ba7-b3a6-3e8172306408\") " pod="openstack/nova-cell0-conductor-db-sync-jdtxs" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.662822 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48793e32-107b-4ba7-b3a6-3e8172306408-config-data\") pod \"nova-cell0-conductor-db-sync-jdtxs\" (UID: \"48793e32-107b-4ba7-b3a6-3e8172306408\") " pod="openstack/nova-cell0-conductor-db-sync-jdtxs" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.668919 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "da12db2f-742c-4bf8-8c9e-3ee0868d0c63" (UID: "da12db2f-742c-4bf8-8c9e-3ee0868d0c63"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.685835 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-kube-api-access-v9fxk" (OuterVolumeSpecName: "kube-api-access-v9fxk") pod "da12db2f-742c-4bf8-8c9e-3ee0868d0c63" (UID: "da12db2f-742c-4bf8-8c9e-3ee0868d0c63"). InnerVolumeSpecName "kube-api-access-v9fxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.773829 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "da12db2f-742c-4bf8-8c9e-3ee0868d0c63" (UID: "da12db2f-742c-4bf8-8c9e-3ee0868d0c63"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.774390 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48793e32-107b-4ba7-b3a6-3e8172306408-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jdtxs\" (UID: \"48793e32-107b-4ba7-b3a6-3e8172306408\") " pod="openstack/nova-cell0-conductor-db-sync-jdtxs" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.774699 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48793e32-107b-4ba7-b3a6-3e8172306408-config-data\") pod \"nova-cell0-conductor-db-sync-jdtxs\" (UID: \"48793e32-107b-4ba7-b3a6-3e8172306408\") " pod="openstack/nova-cell0-conductor-db-sync-jdtxs" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.774934 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzgvx\" (UniqueName: \"kubernetes.io/projected/48793e32-107b-4ba7-b3a6-3e8172306408-kube-api-access-nzgvx\") pod \"nova-cell0-conductor-db-sync-jdtxs\" (UID: \"48793e32-107b-4ba7-b3a6-3e8172306408\") " pod="openstack/nova-cell0-conductor-db-sync-jdtxs" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.775763 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48793e32-107b-4ba7-b3a6-3e8172306408-scripts\") pod \"nova-cell0-conductor-db-sync-jdtxs\" (UID: \"48793e32-107b-4ba7-b3a6-3e8172306408\") " pod="openstack/nova-cell0-conductor-db-sync-jdtxs" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.776091 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9fxk\" (UniqueName: \"kubernetes.io/projected/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-kube-api-access-v9fxk\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.776209 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.776238 4813 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-httpd-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.778859 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48793e32-107b-4ba7-b3a6-3e8172306408-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-jdtxs\" (UID: \"48793e32-107b-4ba7-b3a6-3e8172306408\") " pod="openstack/nova-cell0-conductor-db-sync-jdtxs" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.783451 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48793e32-107b-4ba7-b3a6-3e8172306408-scripts\") pod \"nova-cell0-conductor-db-sync-jdtxs\" (UID: \"48793e32-107b-4ba7-b3a6-3e8172306408\") " pod="openstack/nova-cell0-conductor-db-sync-jdtxs" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.783546 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48793e32-107b-4ba7-b3a6-3e8172306408-config-data\") pod \"nova-cell0-conductor-db-sync-jdtxs\" (UID: \"48793e32-107b-4ba7-b3a6-3e8172306408\") " pod="openstack/nova-cell0-conductor-db-sync-jdtxs" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.784079 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0a52b24b-cc19-4720-a05b-961015362769" path="/var/lib/kubelet/pods/0a52b24b-cc19-4720-a05b-961015362769/volumes" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.802757 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzgvx\" (UniqueName: \"kubernetes.io/projected/48793e32-107b-4ba7-b3a6-3e8172306408-kube-api-access-nzgvx\") pod \"nova-cell0-conductor-db-sync-jdtxs\" (UID: \"48793e32-107b-4ba7-b3a6-3e8172306408\") " pod="openstack/nova-cell0-conductor-db-sync-jdtxs" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.810539 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-config" (OuterVolumeSpecName: "config") pod "da12db2f-742c-4bf8-8c9e-3ee0868d0c63" (UID: "da12db2f-742c-4bf8-8c9e-3ee0868d0c63"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.814244 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "da12db2f-742c-4bf8-8c9e-3ee0868d0c63" (UID: "da12db2f-742c-4bf8-8c9e-3ee0868d0c63"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.819534 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jdtxs" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.883490 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.883808 4813 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/da12db2f-742c-4bf8-8c9e-3ee0868d0c63-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.885496 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.985162 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j75tr\" (UniqueName: \"kubernetes.io/projected/92f6c71b-18e8-4015-a231-a741629a4711-kube-api-access-j75tr\") pod \"92f6c71b-18e8-4015-a231-a741629a4711\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.985204 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-config-data\") pod \"92f6c71b-18e8-4015-a231-a741629a4711\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.985276 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92f6c71b-18e8-4015-a231-a741629a4711-log-httpd\") pod \"92f6c71b-18e8-4015-a231-a741629a4711\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.985297 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92f6c71b-18e8-4015-a231-a741629a4711-run-httpd\") pod \"92f6c71b-18e8-4015-a231-a741629a4711\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.985313 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-combined-ca-bundle\") pod \"92f6c71b-18e8-4015-a231-a741629a4711\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.985373 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-scripts\") pod \"92f6c71b-18e8-4015-a231-a741629a4711\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.985414 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-sg-core-conf-yaml\") pod \"92f6c71b-18e8-4015-a231-a741629a4711\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.986295 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92f6c71b-18e8-4015-a231-a741629a4711-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "92f6c71b-18e8-4015-a231-a741629a4711" (UID: "92f6c71b-18e8-4015-a231-a741629a4711"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.986522 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92f6c71b-18e8-4015-a231-a741629a4711-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "92f6c71b-18e8-4015-a231-a741629a4711" (UID: "92f6c71b-18e8-4015-a231-a741629a4711"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.989392 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92f6c71b-18e8-4015-a231-a741629a4711-kube-api-access-j75tr" (OuterVolumeSpecName: "kube-api-access-j75tr") pod "92f6c71b-18e8-4015-a231-a741629a4711" (UID: "92f6c71b-18e8-4015-a231-a741629a4711"). InnerVolumeSpecName "kube-api-access-j75tr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:32:26 crc kubenswrapper[4813]: I0317 09:32:26.990061 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-scripts" (OuterVolumeSpecName: "scripts") pod "92f6c71b-18e8-4015-a231-a741629a4711" (UID: "92f6c71b-18e8-4015-a231-a741629a4711"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.060650 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "92f6c71b-18e8-4015-a231-a741629a4711" (UID: "92f6c71b-18e8-4015-a231-a741629a4711"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.086412 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "92f6c71b-18e8-4015-a231-a741629a4711" (UID: "92f6c71b-18e8-4015-a231-a741629a4711"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.087128 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-combined-ca-bundle\") pod \"92f6c71b-18e8-4015-a231-a741629a4711\" (UID: \"92f6c71b-18e8-4015-a231-a741629a4711\") " Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.087658 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.087675 4813 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.087686 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j75tr\" (UniqueName: \"kubernetes.io/projected/92f6c71b-18e8-4015-a231-a741629a4711-kube-api-access-j75tr\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.087697 4813 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92f6c71b-18e8-4015-a231-a741629a4711-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.087706 4813 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/92f6c71b-18e8-4015-a231-a741629a4711-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:27 crc kubenswrapper[4813]: W0317 09:32:27.087782 4813 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/92f6c71b-18e8-4015-a231-a741629a4711/volumes/kubernetes.io~secret/combined-ca-bundle Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.087795 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "92f6c71b-18e8-4015-a231-a741629a4711" (UID: "92f6c71b-18e8-4015-a231-a741629a4711"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.118230 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-config-data" (OuterVolumeSpecName: "config-data") pod "92f6c71b-18e8-4015-a231-a741629a4711" (UID: "92f6c71b-18e8-4015-a231-a741629a4711"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.178418 4813 generic.go:334] "Generic (PLEG): container finished" podID="92f6c71b-18e8-4015-a231-a741629a4711" containerID="4ccef2d8b50e05342b845d860d7b1a39b82ccca0fd483c11ef1958029df47cfd" exitCode=0 Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.178489 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92f6c71b-18e8-4015-a231-a741629a4711","Type":"ContainerDied","Data":"4ccef2d8b50e05342b845d860d7b1a39b82ccca0fd483c11ef1958029df47cfd"} Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.178519 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"92f6c71b-18e8-4015-a231-a741629a4711","Type":"ContainerDied","Data":"d386ba7487283c309d72f0e6c93fd1891a38efa8f8fc1ffec822391cf60f2d28"} Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.178537 4813 scope.go:117] "RemoveContainer" containerID="44163684427cfcdba1b739c304593eb2d1a04589d09d1b0e2ca1127abcd26d4b" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.178565 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.182030 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-b5495b5bd-g77kp" event={"ID":"da12db2f-742c-4bf8-8c9e-3ee0868d0c63","Type":"ContainerDied","Data":"bf28a2ce3451de4a157ee9315e929006bac898fa25cde725ee6f38f9ecec6311"} Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.182114 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-b5495b5bd-g77kp" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.188908 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.188934 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92f6c71b-18e8-4015-a231-a741629a4711-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.207663 4813 scope.go:117] "RemoveContainer" containerID="e39df0ae8b176e5ec9935f93b5a9c0f1d1bcf2895a38dfca5b4e67f9beea5969" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.242313 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.242531 4813 scope.go:117] "RemoveContainer" containerID="2eb3949ddb871cc652a4e28ce50e6e40979688ec8a5a9d97a0ac2518754b7e3b" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.270670 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.300347 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-b5495b5bd-g77kp"] Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.320365 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-b5495b5bd-g77kp"] Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.320378 4813 scope.go:117] "RemoveContainer" containerID="4ccef2d8b50e05342b845d860d7b1a39b82ccca0fd483c11ef1958029df47cfd" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.329719 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:32:27 crc kubenswrapper[4813]: E0317 09:32:27.330050 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92f6c71b-18e8-4015-a231-a741629a4711" containerName="sg-core" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.330064 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="92f6c71b-18e8-4015-a231-a741629a4711" containerName="sg-core" Mar 17 09:32:27 crc kubenswrapper[4813]: E0317 09:32:27.330071 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da12db2f-742c-4bf8-8c9e-3ee0868d0c63" containerName="neutron-httpd" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.330077 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="da12db2f-742c-4bf8-8c9e-3ee0868d0c63" containerName="neutron-httpd" Mar 17 09:32:27 crc kubenswrapper[4813]: E0317 09:32:27.330086 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92f6c71b-18e8-4015-a231-a741629a4711" containerName="ceilometer-central-agent" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.330092 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="92f6c71b-18e8-4015-a231-a741629a4711" containerName="ceilometer-central-agent" Mar 17 09:32:27 crc kubenswrapper[4813]: E0317 09:32:27.330118 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da12db2f-742c-4bf8-8c9e-3ee0868d0c63" containerName="neutron-api" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.330123 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="da12db2f-742c-4bf8-8c9e-3ee0868d0c63" containerName="neutron-api" Mar 17 09:32:27 crc kubenswrapper[4813]: E0317 09:32:27.330134 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92f6c71b-18e8-4015-a231-a741629a4711" containerName="proxy-httpd" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.330139 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="92f6c71b-18e8-4015-a231-a741629a4711" containerName="proxy-httpd" Mar 17 09:32:27 crc kubenswrapper[4813]: E0317 09:32:27.330153 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92f6c71b-18e8-4015-a231-a741629a4711" containerName="ceilometer-notification-agent" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.330159 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="92f6c71b-18e8-4015-a231-a741629a4711" containerName="ceilometer-notification-agent" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.330314 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="92f6c71b-18e8-4015-a231-a741629a4711" containerName="ceilometer-central-agent" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.330328 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="da12db2f-742c-4bf8-8c9e-3ee0868d0c63" containerName="neutron-api" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.330344 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="92f6c71b-18e8-4015-a231-a741629a4711" containerName="ceilometer-notification-agent" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.330356 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="da12db2f-742c-4bf8-8c9e-3ee0868d0c63" containerName="neutron-httpd" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.330365 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="92f6c71b-18e8-4015-a231-a741629a4711" containerName="sg-core" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.330375 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="92f6c71b-18e8-4015-a231-a741629a4711" containerName="proxy-httpd" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.331793 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.334135 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.334363 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.342012 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.342070 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.343223 4813 scope.go:117] "RemoveContainer" containerID="44163684427cfcdba1b739c304593eb2d1a04589d09d1b0e2ca1127abcd26d4b" Mar 17 09:32:27 crc kubenswrapper[4813]: E0317 09:32:27.346323 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44163684427cfcdba1b739c304593eb2d1a04589d09d1b0e2ca1127abcd26d4b\": container with ID starting with 44163684427cfcdba1b739c304593eb2d1a04589d09d1b0e2ca1127abcd26d4b not found: ID does not exist" containerID="44163684427cfcdba1b739c304593eb2d1a04589d09d1b0e2ca1127abcd26d4b" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.346380 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44163684427cfcdba1b739c304593eb2d1a04589d09d1b0e2ca1127abcd26d4b"} err="failed to get container status \"44163684427cfcdba1b739c304593eb2d1a04589d09d1b0e2ca1127abcd26d4b\": rpc error: code = NotFound desc = could not find container \"44163684427cfcdba1b739c304593eb2d1a04589d09d1b0e2ca1127abcd26d4b\": container with ID starting with 44163684427cfcdba1b739c304593eb2d1a04589d09d1b0e2ca1127abcd26d4b not found: ID does not exist" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.346416 4813 scope.go:117] "RemoveContainer" containerID="e39df0ae8b176e5ec9935f93b5a9c0f1d1bcf2895a38dfca5b4e67f9beea5969" Mar 17 09:32:27 crc kubenswrapper[4813]: E0317 09:32:27.346856 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e39df0ae8b176e5ec9935f93b5a9c0f1d1bcf2895a38dfca5b4e67f9beea5969\": container with ID starting with e39df0ae8b176e5ec9935f93b5a9c0f1d1bcf2895a38dfca5b4e67f9beea5969 not found: ID does not exist" containerID="e39df0ae8b176e5ec9935f93b5a9c0f1d1bcf2895a38dfca5b4e67f9beea5969" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.346886 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e39df0ae8b176e5ec9935f93b5a9c0f1d1bcf2895a38dfca5b4e67f9beea5969"} err="failed to get container status \"e39df0ae8b176e5ec9935f93b5a9c0f1d1bcf2895a38dfca5b4e67f9beea5969\": rpc error: code = NotFound desc = could not find container \"e39df0ae8b176e5ec9935f93b5a9c0f1d1bcf2895a38dfca5b4e67f9beea5969\": container with ID starting with e39df0ae8b176e5ec9935f93b5a9c0f1d1bcf2895a38dfca5b4e67f9beea5969 not found: ID does not exist" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.346909 4813 scope.go:117] "RemoveContainer" containerID="2eb3949ddb871cc652a4e28ce50e6e40979688ec8a5a9d97a0ac2518754b7e3b" Mar 17 09:32:27 crc kubenswrapper[4813]: E0317 09:32:27.347247 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2eb3949ddb871cc652a4e28ce50e6e40979688ec8a5a9d97a0ac2518754b7e3b\": container with ID starting with 2eb3949ddb871cc652a4e28ce50e6e40979688ec8a5a9d97a0ac2518754b7e3b not found: ID does not exist" containerID="2eb3949ddb871cc652a4e28ce50e6e40979688ec8a5a9d97a0ac2518754b7e3b" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.347264 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2eb3949ddb871cc652a4e28ce50e6e40979688ec8a5a9d97a0ac2518754b7e3b"} err="failed to get container status \"2eb3949ddb871cc652a4e28ce50e6e40979688ec8a5a9d97a0ac2518754b7e3b\": rpc error: code = NotFound desc = could not find container \"2eb3949ddb871cc652a4e28ce50e6e40979688ec8a5a9d97a0ac2518754b7e3b\": container with ID starting with 2eb3949ddb871cc652a4e28ce50e6e40979688ec8a5a9d97a0ac2518754b7e3b not found: ID does not exist" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.347275 4813 scope.go:117] "RemoveContainer" containerID="4ccef2d8b50e05342b845d860d7b1a39b82ccca0fd483c11ef1958029df47cfd" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.347323 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:32:27 crc kubenswrapper[4813]: E0317 09:32:27.347727 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ccef2d8b50e05342b845d860d7b1a39b82ccca0fd483c11ef1958029df47cfd\": container with ID starting with 4ccef2d8b50e05342b845d860d7b1a39b82ccca0fd483c11ef1958029df47cfd not found: ID does not exist" containerID="4ccef2d8b50e05342b845d860d7b1a39b82ccca0fd483c11ef1958029df47cfd" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.347745 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ccef2d8b50e05342b845d860d7b1a39b82ccca0fd483c11ef1958029df47cfd"} err="failed to get container status \"4ccef2d8b50e05342b845d860d7b1a39b82ccca0fd483c11ef1958029df47cfd\": rpc error: code = NotFound desc = could not find container \"4ccef2d8b50e05342b845d860d7b1a39b82ccca0fd483c11ef1958029df47cfd\": container with ID starting with 4ccef2d8b50e05342b845d860d7b1a39b82ccca0fd483c11ef1958029df47cfd not found: ID does not exist" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.347758 4813 scope.go:117] "RemoveContainer" containerID="c84fbe3bc0c39f7fff03f9ed586e1ddcb0e72ef85170e94045a5c1c7eabcb226" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.354876 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jdtxs"] Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.383121 4813 scope.go:117] "RemoveContainer" containerID="8c5f1c181c9c83af6c9eac398fcc89f4f7500de195959ec00e013725d482ce66" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.394364 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.394446 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8efe575c-ec1e-4678-8c52-db4dd272d4d1-run-httpd\") pod \"ceilometer-0\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.394734 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.394832 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8efe575c-ec1e-4678-8c52-db4dd272d4d1-log-httpd\") pod \"ceilometer-0\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.394895 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4r7xv\" (UniqueName: \"kubernetes.io/projected/8efe575c-ec1e-4678-8c52-db4dd272d4d1-kube-api-access-4r7xv\") pod \"ceilometer-0\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.395063 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-config-data\") pod \"ceilometer-0\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.395161 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-scripts\") pod \"ceilometer-0\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.395588 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.396082 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.495834 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.495878 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8efe575c-ec1e-4678-8c52-db4dd272d4d1-run-httpd\") pod \"ceilometer-0\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.495921 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.495965 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8efe575c-ec1e-4678-8c52-db4dd272d4d1-log-httpd\") pod \"ceilometer-0\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.495991 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4r7xv\" (UniqueName: \"kubernetes.io/projected/8efe575c-ec1e-4678-8c52-db4dd272d4d1-kube-api-access-4r7xv\") pod \"ceilometer-0\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.496024 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-config-data\") pod \"ceilometer-0\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.496052 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-scripts\") pod \"ceilometer-0\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.496562 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8efe575c-ec1e-4678-8c52-db4dd272d4d1-run-httpd\") pod \"ceilometer-0\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.496617 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8efe575c-ec1e-4678-8c52-db4dd272d4d1-log-httpd\") pod \"ceilometer-0\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.500783 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.501872 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-scripts\") pod \"ceilometer-0\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.502563 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-config-data\") pod \"ceilometer-0\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.502883 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.518760 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4r7xv\" (UniqueName: \"kubernetes.io/projected/8efe575c-ec1e-4678-8c52-db4dd272d4d1-kube-api-access-4r7xv\") pod \"ceilometer-0\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " pod="openstack/ceilometer-0" Mar 17 09:32:27 crc kubenswrapper[4813]: I0317 09:32:27.683881 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:32:28 crc kubenswrapper[4813]: I0317 09:32:28.199887 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jdtxs" event={"ID":"48793e32-107b-4ba7-b3a6-3e8172306408","Type":"ContainerStarted","Data":"21fca64e599b1bac11a68ee1c8f71233c7ddc665382163caf82a92a51bfb5408"} Mar 17 09:32:28 crc kubenswrapper[4813]: I0317 09:32:28.200220 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 17 09:32:28 crc kubenswrapper[4813]: I0317 09:32:28.200234 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Mar 17 09:32:28 crc kubenswrapper[4813]: I0317 09:32:28.278317 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:32:28 crc kubenswrapper[4813]: I0317 09:32:28.480913 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:32:28 crc kubenswrapper[4813]: I0317 09:32:28.753753 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92f6c71b-18e8-4015-a231-a741629a4711" path="/var/lib/kubelet/pods/92f6c71b-18e8-4015-a231-a741629a4711/volumes" Mar 17 09:32:28 crc kubenswrapper[4813]: I0317 09:32:28.757731 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da12db2f-742c-4bf8-8c9e-3ee0868d0c63" path="/var/lib/kubelet/pods/da12db2f-742c-4bf8-8c9e-3ee0868d0c63/volumes" Mar 17 09:32:29 crc kubenswrapper[4813]: I0317 09:32:29.256484 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8efe575c-ec1e-4678-8c52-db4dd272d4d1","Type":"ContainerStarted","Data":"448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59"} Mar 17 09:32:29 crc kubenswrapper[4813]: I0317 09:32:29.256754 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8efe575c-ec1e-4678-8c52-db4dd272d4d1","Type":"ContainerStarted","Data":"43e369f7773f8ff7bc36e60b6b39ace63a6c503b7c4c14145c0596e4d6c3baa2"} Mar 17 09:32:29 crc kubenswrapper[4813]: I0317 09:32:29.745095 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-557566c676-bbf8q" Mar 17 09:32:29 crc kubenswrapper[4813]: I0317 09:32:29.773962 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-557566c676-bbf8q" Mar 17 09:32:29 crc kubenswrapper[4813]: I0317 09:32:29.865981 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-6cc74cf554-dmnmf"] Mar 17 09:32:29 crc kubenswrapper[4813]: I0317 09:32:29.866328 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-6cc74cf554-dmnmf" podUID="bb24e4e7-5911-4a4b-bff0-e0de7087be65" containerName="placement-log" containerID="cri-o://0df61a5ea2fb36b21335d446d9790e9116aa82a0d3dfb4105c46092354b3a6d2" gracePeriod=30 Mar 17 09:32:29 crc kubenswrapper[4813]: I0317 09:32:29.866436 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/placement-6cc74cf554-dmnmf" podUID="bb24e4e7-5911-4a4b-bff0-e0de7087be65" containerName="placement-api" containerID="cri-o://f54d815fe296100a77eea0b4c0830a67b317b2b603abccbc9fa035e4ca28c4bb" gracePeriod=30 Mar 17 09:32:30 crc kubenswrapper[4813]: I0317 09:32:30.269000 4813 generic.go:334] "Generic (PLEG): container finished" podID="bb24e4e7-5911-4a4b-bff0-e0de7087be65" containerID="0df61a5ea2fb36b21335d446d9790e9116aa82a0d3dfb4105c46092354b3a6d2" exitCode=143 Mar 17 09:32:30 crc kubenswrapper[4813]: I0317 09:32:30.269150 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6cc74cf554-dmnmf" event={"ID":"bb24e4e7-5911-4a4b-bff0-e0de7087be65","Type":"ContainerDied","Data":"0df61a5ea2fb36b21335d446d9790e9116aa82a0d3dfb4105c46092354b3a6d2"} Mar 17 09:32:30 crc kubenswrapper[4813]: I0317 09:32:30.272151 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8efe575c-ec1e-4678-8c52-db4dd272d4d1","Type":"ContainerStarted","Data":"c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889"} Mar 17 09:32:30 crc kubenswrapper[4813]: I0317 09:32:30.364720 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 17 09:32:30 crc kubenswrapper[4813]: I0317 09:32:30.364768 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Mar 17 09:32:30 crc kubenswrapper[4813]: I0317 09:32:30.407684 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 17 09:32:30 crc kubenswrapper[4813]: I0317 09:32:30.411830 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Mar 17 09:32:30 crc kubenswrapper[4813]: I0317 09:32:30.868073 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 17 09:32:30 crc kubenswrapper[4813]: I0317 09:32:30.868460 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 17 09:32:30 crc kubenswrapper[4813]: I0317 09:32:30.870454 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Mar 17 09:32:31 crc kubenswrapper[4813]: I0317 09:32:31.293575 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8efe575c-ec1e-4678-8c52-db4dd272d4d1","Type":"ContainerStarted","Data":"3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544"} Mar 17 09:32:31 crc kubenswrapper[4813]: I0317 09:32:31.293952 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 17 09:32:31 crc kubenswrapper[4813]: I0317 09:32:31.293983 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Mar 17 09:32:33 crc kubenswrapper[4813]: I0317 09:32:33.317483 4813 generic.go:334] "Generic (PLEG): container finished" podID="bb24e4e7-5911-4a4b-bff0-e0de7087be65" containerID="f54d815fe296100a77eea0b4c0830a67b317b2b603abccbc9fa035e4ca28c4bb" exitCode=0 Mar 17 09:32:33 crc kubenswrapper[4813]: I0317 09:32:33.317553 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6cc74cf554-dmnmf" event={"ID":"bb24e4e7-5911-4a4b-bff0-e0de7087be65","Type":"ContainerDied","Data":"f54d815fe296100a77eea0b4c0830a67b317b2b603abccbc9fa035e4ca28c4bb"} Mar 17 09:32:33 crc kubenswrapper[4813]: I0317 09:32:33.800676 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 17 09:32:33 crc kubenswrapper[4813]: I0317 09:32:33.800815 4813 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Mar 17 09:32:33 crc kubenswrapper[4813]: I0317 09:32:33.805012 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.215539 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.302809 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-internal-tls-certs\") pod \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.302881 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbb9\" (UniqueName: \"kubernetes.io/projected/bb24e4e7-5911-4a4b-bff0-e0de7087be65-kube-api-access-jhbb9\") pod \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.302962 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb24e4e7-5911-4a4b-bff0-e0de7087be65-logs\") pod \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.303224 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-combined-ca-bundle\") pod \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.303356 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-config-data\") pod \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.303476 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-scripts\") pod \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.303650 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-public-tls-certs\") pod \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\" (UID: \"bb24e4e7-5911-4a4b-bff0-e0de7087be65\") " Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.303689 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb24e4e7-5911-4a4b-bff0-e0de7087be65-logs" (OuterVolumeSpecName: "logs") pod "bb24e4e7-5911-4a4b-bff0-e0de7087be65" (UID: "bb24e4e7-5911-4a4b-bff0-e0de7087be65"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.304425 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb24e4e7-5911-4a4b-bff0-e0de7087be65-logs\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.310082 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-scripts" (OuterVolumeSpecName: "scripts") pod "bb24e4e7-5911-4a4b-bff0-e0de7087be65" (UID: "bb24e4e7-5911-4a4b-bff0-e0de7087be65"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.314345 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb24e4e7-5911-4a4b-bff0-e0de7087be65-kube-api-access-jhbb9" (OuterVolumeSpecName: "kube-api-access-jhbb9") pod "bb24e4e7-5911-4a4b-bff0-e0de7087be65" (UID: "bb24e4e7-5911-4a4b-bff0-e0de7087be65"). InnerVolumeSpecName "kube-api-access-jhbb9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.366889 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-6cc74cf554-dmnmf" event={"ID":"bb24e4e7-5911-4a4b-bff0-e0de7087be65","Type":"ContainerDied","Data":"801f34584e1cf811b7cda08ea8a4678f0f527fd3d3b37b10c0ae9e0211c56381"} Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.366948 4813 scope.go:117] "RemoveContainer" containerID="f54d815fe296100a77eea0b4c0830a67b317b2b603abccbc9fa035e4ca28c4bb" Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.366957 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-6cc74cf554-dmnmf" Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.392050 4813 scope.go:117] "RemoveContainer" containerID="0df61a5ea2fb36b21335d446d9790e9116aa82a0d3dfb4105c46092354b3a6d2" Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.406671 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbb9\" (UniqueName: \"kubernetes.io/projected/bb24e4e7-5911-4a4b-bff0-e0de7087be65-kube-api-access-jhbb9\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.406697 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.422753 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-config-data" (OuterVolumeSpecName: "config-data") pod "bb24e4e7-5911-4a4b-bff0-e0de7087be65" (UID: "bb24e4e7-5911-4a4b-bff0-e0de7087be65"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.434724 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb24e4e7-5911-4a4b-bff0-e0de7087be65" (UID: "bb24e4e7-5911-4a4b-bff0-e0de7087be65"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.483231 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "bb24e4e7-5911-4a4b-bff0-e0de7087be65" (UID: "bb24e4e7-5911-4a4b-bff0-e0de7087be65"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.493527 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "bb24e4e7-5911-4a4b-bff0-e0de7087be65" (UID: "bb24e4e7-5911-4a4b-bff0-e0de7087be65"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.508817 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.508855 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.508871 4813 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.508883 4813 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb24e4e7-5911-4a4b-bff0-e0de7087be65-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.707017 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-6cc74cf554-dmnmf"] Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.718661 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-6cc74cf554-dmnmf"] Mar 17 09:32:38 crc kubenswrapper[4813]: I0317 09:32:38.743477 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb24e4e7-5911-4a4b-bff0-e0de7087be65" path="/var/lib/kubelet/pods/bb24e4e7-5911-4a4b-bff0-e0de7087be65/volumes" Mar 17 09:32:39 crc kubenswrapper[4813]: I0317 09:32:39.386482 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jdtxs" event={"ID":"48793e32-107b-4ba7-b3a6-3e8172306408","Type":"ContainerStarted","Data":"cd051cb78eab1e8bb6bdda3cac209d8aa12d7ecdf22cc87948d4018db9a49829"} Mar 17 09:32:39 crc kubenswrapper[4813]: I0317 09:32:39.390032 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8efe575c-ec1e-4678-8c52-db4dd272d4d1","Type":"ContainerStarted","Data":"6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f"} Mar 17 09:32:39 crc kubenswrapper[4813]: I0317 09:32:39.390269 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8efe575c-ec1e-4678-8c52-db4dd272d4d1" containerName="ceilometer-central-agent" containerID="cri-o://448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59" gracePeriod=30 Mar 17 09:32:39 crc kubenswrapper[4813]: I0317 09:32:39.390413 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 17 09:32:39 crc kubenswrapper[4813]: I0317 09:32:39.390430 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8efe575c-ec1e-4678-8c52-db4dd272d4d1" containerName="proxy-httpd" containerID="cri-o://6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f" gracePeriod=30 Mar 17 09:32:39 crc kubenswrapper[4813]: I0317 09:32:39.390474 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8efe575c-ec1e-4678-8c52-db4dd272d4d1" containerName="ceilometer-notification-agent" containerID="cri-o://c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889" gracePeriod=30 Mar 17 09:32:39 crc kubenswrapper[4813]: I0317 09:32:39.390546 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8efe575c-ec1e-4678-8c52-db4dd272d4d1" containerName="sg-core" containerID="cri-o://3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544" gracePeriod=30 Mar 17 09:32:39 crc kubenswrapper[4813]: I0317 09:32:39.427620 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-jdtxs" podStartSLOduration=2.74726041 podStartE2EDuration="13.427590644s" podCreationTimestamp="2026-03-17 09:32:26 +0000 UTC" firstStartedPulling="2026-03-17 09:32:27.343324974 +0000 UTC m=+1369.444128473" lastFinishedPulling="2026-03-17 09:32:38.023655208 +0000 UTC m=+1380.124458707" observedRunningTime="2026-03-17 09:32:39.421292785 +0000 UTC m=+1381.522096294" watchObservedRunningTime="2026-03-17 09:32:39.427590644 +0000 UTC m=+1381.528394133" Mar 17 09:32:39 crc kubenswrapper[4813]: I0317 09:32:39.450683 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.780375321 podStartE2EDuration="12.450645158s" podCreationTimestamp="2026-03-17 09:32:27 +0000 UTC" firstStartedPulling="2026-03-17 09:32:28.335197129 +0000 UTC m=+1370.436000628" lastFinishedPulling="2026-03-17 09:32:38.005466956 +0000 UTC m=+1380.106270465" observedRunningTime="2026-03-17 09:32:39.446307672 +0000 UTC m=+1381.547111171" watchObservedRunningTime="2026-03-17 09:32:39.450645158 +0000 UTC m=+1381.551448657" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.283332 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.347691 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-config-data\") pod \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.347829 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8efe575c-ec1e-4678-8c52-db4dd272d4d1-run-httpd\") pod \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.347876 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-sg-core-conf-yaml\") pod \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.347935 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8efe575c-ec1e-4678-8c52-db4dd272d4d1-log-httpd\") pod \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.347983 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4r7xv\" (UniqueName: \"kubernetes.io/projected/8efe575c-ec1e-4678-8c52-db4dd272d4d1-kube-api-access-4r7xv\") pod \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.348007 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-combined-ca-bundle\") pod \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.348040 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-scripts\") pod \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\" (UID: \"8efe575c-ec1e-4678-8c52-db4dd272d4d1\") " Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.348814 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8efe575c-ec1e-4678-8c52-db4dd272d4d1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8efe575c-ec1e-4678-8c52-db4dd272d4d1" (UID: "8efe575c-ec1e-4678-8c52-db4dd272d4d1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.349092 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8efe575c-ec1e-4678-8c52-db4dd272d4d1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8efe575c-ec1e-4678-8c52-db4dd272d4d1" (UID: "8efe575c-ec1e-4678-8c52-db4dd272d4d1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.353391 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8efe575c-ec1e-4678-8c52-db4dd272d4d1-kube-api-access-4r7xv" (OuterVolumeSpecName: "kube-api-access-4r7xv") pod "8efe575c-ec1e-4678-8c52-db4dd272d4d1" (UID: "8efe575c-ec1e-4678-8c52-db4dd272d4d1"). InnerVolumeSpecName "kube-api-access-4r7xv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.353399 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-scripts" (OuterVolumeSpecName: "scripts") pod "8efe575c-ec1e-4678-8c52-db4dd272d4d1" (UID: "8efe575c-ec1e-4678-8c52-db4dd272d4d1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.396481 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8efe575c-ec1e-4678-8c52-db4dd272d4d1" (UID: "8efe575c-ec1e-4678-8c52-db4dd272d4d1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.401114 4813 generic.go:334] "Generic (PLEG): container finished" podID="8efe575c-ec1e-4678-8c52-db4dd272d4d1" containerID="6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f" exitCode=0 Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.401164 4813 generic.go:334] "Generic (PLEG): container finished" podID="8efe575c-ec1e-4678-8c52-db4dd272d4d1" containerID="3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544" exitCode=2 Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.401172 4813 generic.go:334] "Generic (PLEG): container finished" podID="8efe575c-ec1e-4678-8c52-db4dd272d4d1" containerID="c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889" exitCode=0 Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.401179 4813 generic.go:334] "Generic (PLEG): container finished" podID="8efe575c-ec1e-4678-8c52-db4dd272d4d1" containerID="448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59" exitCode=0 Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.401192 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.401191 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8efe575c-ec1e-4678-8c52-db4dd272d4d1","Type":"ContainerDied","Data":"6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f"} Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.401236 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8efe575c-ec1e-4678-8c52-db4dd272d4d1","Type":"ContainerDied","Data":"3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544"} Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.401247 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8efe575c-ec1e-4678-8c52-db4dd272d4d1","Type":"ContainerDied","Data":"c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889"} Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.401256 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8efe575c-ec1e-4678-8c52-db4dd272d4d1","Type":"ContainerDied","Data":"448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59"} Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.401264 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8efe575c-ec1e-4678-8c52-db4dd272d4d1","Type":"ContainerDied","Data":"43e369f7773f8ff7bc36e60b6b39ace63a6c503b7c4c14145c0596e4d6c3baa2"} Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.401278 4813 scope.go:117] "RemoveContainer" containerID="6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.424593 4813 scope.go:117] "RemoveContainer" containerID="3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.434882 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8efe575c-ec1e-4678-8c52-db4dd272d4d1" (UID: "8efe575c-ec1e-4678-8c52-db4dd272d4d1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.445338 4813 scope.go:117] "RemoveContainer" containerID="c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.449631 4813 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8efe575c-ec1e-4678-8c52-db4dd272d4d1-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.449653 4813 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.449662 4813 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8efe575c-ec1e-4678-8c52-db4dd272d4d1-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.449672 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4r7xv\" (UniqueName: \"kubernetes.io/projected/8efe575c-ec1e-4678-8c52-db4dd272d4d1-kube-api-access-4r7xv\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.449681 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.449688 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.463893 4813 scope.go:117] "RemoveContainer" containerID="448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.463997 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-config-data" (OuterVolumeSpecName: "config-data") pod "8efe575c-ec1e-4678-8c52-db4dd272d4d1" (UID: "8efe575c-ec1e-4678-8c52-db4dd272d4d1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.486723 4813 scope.go:117] "RemoveContainer" containerID="6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f" Mar 17 09:32:40 crc kubenswrapper[4813]: E0317 09:32:40.487282 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f\": container with ID starting with 6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f not found: ID does not exist" containerID="6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.487328 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f"} err="failed to get container status \"6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f\": rpc error: code = NotFound desc = could not find container \"6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f\": container with ID starting with 6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f not found: ID does not exist" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.487358 4813 scope.go:117] "RemoveContainer" containerID="3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544" Mar 17 09:32:40 crc kubenswrapper[4813]: E0317 09:32:40.487816 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544\": container with ID starting with 3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544 not found: ID does not exist" containerID="3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.487862 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544"} err="failed to get container status \"3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544\": rpc error: code = NotFound desc = could not find container \"3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544\": container with ID starting with 3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544 not found: ID does not exist" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.487888 4813 scope.go:117] "RemoveContainer" containerID="c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889" Mar 17 09:32:40 crc kubenswrapper[4813]: E0317 09:32:40.488240 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889\": container with ID starting with c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889 not found: ID does not exist" containerID="c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.488274 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889"} err="failed to get container status \"c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889\": rpc error: code = NotFound desc = could not find container \"c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889\": container with ID starting with c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889 not found: ID does not exist" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.488294 4813 scope.go:117] "RemoveContainer" containerID="448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59" Mar 17 09:32:40 crc kubenswrapper[4813]: E0317 09:32:40.488630 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59\": container with ID starting with 448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59 not found: ID does not exist" containerID="448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.488658 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59"} err="failed to get container status \"448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59\": rpc error: code = NotFound desc = could not find container \"448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59\": container with ID starting with 448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59 not found: ID does not exist" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.488677 4813 scope.go:117] "RemoveContainer" containerID="6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.488984 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f"} err="failed to get container status \"6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f\": rpc error: code = NotFound desc = could not find container \"6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f\": container with ID starting with 6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f not found: ID does not exist" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.489015 4813 scope.go:117] "RemoveContainer" containerID="3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.489264 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544"} err="failed to get container status \"3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544\": rpc error: code = NotFound desc = could not find container \"3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544\": container with ID starting with 3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544 not found: ID does not exist" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.489296 4813 scope.go:117] "RemoveContainer" containerID="c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.489531 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889"} err="failed to get container status \"c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889\": rpc error: code = NotFound desc = could not find container \"c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889\": container with ID starting with c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889 not found: ID does not exist" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.489558 4813 scope.go:117] "RemoveContainer" containerID="448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.489798 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59"} err="failed to get container status \"448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59\": rpc error: code = NotFound desc = could not find container \"448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59\": container with ID starting with 448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59 not found: ID does not exist" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.489821 4813 scope.go:117] "RemoveContainer" containerID="6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.490112 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f"} err="failed to get container status \"6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f\": rpc error: code = NotFound desc = could not find container \"6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f\": container with ID starting with 6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f not found: ID does not exist" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.490132 4813 scope.go:117] "RemoveContainer" containerID="3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.490300 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544"} err="failed to get container status \"3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544\": rpc error: code = NotFound desc = could not find container \"3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544\": container with ID starting with 3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544 not found: ID does not exist" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.490318 4813 scope.go:117] "RemoveContainer" containerID="c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.490477 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889"} err="failed to get container status \"c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889\": rpc error: code = NotFound desc = could not find container \"c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889\": container with ID starting with c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889 not found: ID does not exist" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.490498 4813 scope.go:117] "RemoveContainer" containerID="448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.490695 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59"} err="failed to get container status \"448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59\": rpc error: code = NotFound desc = could not find container \"448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59\": container with ID starting with 448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59 not found: ID does not exist" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.490727 4813 scope.go:117] "RemoveContainer" containerID="6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.490925 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f"} err="failed to get container status \"6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f\": rpc error: code = NotFound desc = could not find container \"6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f\": container with ID starting with 6d25e4bcc519334b6202e6ba9566784de20f1285714fec9b1d2506e485b84a2f not found: ID does not exist" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.490951 4813 scope.go:117] "RemoveContainer" containerID="3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.491230 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544"} err="failed to get container status \"3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544\": rpc error: code = NotFound desc = could not find container \"3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544\": container with ID starting with 3feb9840c7f39d4f4c57b249ea936bf01f3e3fcb5c646bc0e85b9a82b011b544 not found: ID does not exist" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.491255 4813 scope.go:117] "RemoveContainer" containerID="c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.491509 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889"} err="failed to get container status \"c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889\": rpc error: code = NotFound desc = could not find container \"c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889\": container with ID starting with c61438156a7881f01a51dc1e443b5404400fd3b22d18e9902eee2597a0780889 not found: ID does not exist" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.491534 4813 scope.go:117] "RemoveContainer" containerID="448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.491788 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59"} err="failed to get container status \"448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59\": rpc error: code = NotFound desc = could not find container \"448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59\": container with ID starting with 448fc1964a657a6b5d9874428e80e90740f6ba03bae9e5388224a6d01b450a59 not found: ID does not exist" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.551792 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8efe575c-ec1e-4678-8c52-db4dd272d4d1-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.742008 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.742634 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.773778 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:32:40 crc kubenswrapper[4813]: E0317 09:32:40.774253 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb24e4e7-5911-4a4b-bff0-e0de7087be65" containerName="placement-api" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.774276 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb24e4e7-5911-4a4b-bff0-e0de7087be65" containerName="placement-api" Mar 17 09:32:40 crc kubenswrapper[4813]: E0317 09:32:40.774290 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8efe575c-ec1e-4678-8c52-db4dd272d4d1" containerName="proxy-httpd" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.774298 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8efe575c-ec1e-4678-8c52-db4dd272d4d1" containerName="proxy-httpd" Mar 17 09:32:40 crc kubenswrapper[4813]: E0317 09:32:40.774309 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb24e4e7-5911-4a4b-bff0-e0de7087be65" containerName="placement-log" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.774318 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb24e4e7-5911-4a4b-bff0-e0de7087be65" containerName="placement-log" Mar 17 09:32:40 crc kubenswrapper[4813]: E0317 09:32:40.774343 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8efe575c-ec1e-4678-8c52-db4dd272d4d1" containerName="sg-core" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.774352 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8efe575c-ec1e-4678-8c52-db4dd272d4d1" containerName="sg-core" Mar 17 09:32:40 crc kubenswrapper[4813]: E0317 09:32:40.774372 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8efe575c-ec1e-4678-8c52-db4dd272d4d1" containerName="ceilometer-central-agent" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.774380 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8efe575c-ec1e-4678-8c52-db4dd272d4d1" containerName="ceilometer-central-agent" Mar 17 09:32:40 crc kubenswrapper[4813]: E0317 09:32:40.774400 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8efe575c-ec1e-4678-8c52-db4dd272d4d1" containerName="ceilometer-notification-agent" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.774407 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8efe575c-ec1e-4678-8c52-db4dd272d4d1" containerName="ceilometer-notification-agent" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.774636 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb24e4e7-5911-4a4b-bff0-e0de7087be65" containerName="placement-log" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.774649 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8efe575c-ec1e-4678-8c52-db4dd272d4d1" containerName="sg-core" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.774663 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8efe575c-ec1e-4678-8c52-db4dd272d4d1" containerName="ceilometer-notification-agent" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.774687 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8efe575c-ec1e-4678-8c52-db4dd272d4d1" containerName="proxy-httpd" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.774696 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8efe575c-ec1e-4678-8c52-db4dd272d4d1" containerName="ceilometer-central-agent" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.774708 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb24e4e7-5911-4a4b-bff0-e0de7087be65" containerName="placement-api" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.776790 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.779418 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.779824 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.785655 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.856891 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f41fef51-25cf-4306-87fa-4f203d7d18e9-run-httpd\") pod \"ceilometer-0\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.857029 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jtjn\" (UniqueName: \"kubernetes.io/projected/f41fef51-25cf-4306-87fa-4f203d7d18e9-kube-api-access-2jtjn\") pod \"ceilometer-0\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.857080 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-scripts\") pod \"ceilometer-0\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.857230 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-config-data\") pod \"ceilometer-0\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.857313 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f41fef51-25cf-4306-87fa-4f203d7d18e9-log-httpd\") pod \"ceilometer-0\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.857369 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.857445 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.959084 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jtjn\" (UniqueName: \"kubernetes.io/projected/f41fef51-25cf-4306-87fa-4f203d7d18e9-kube-api-access-2jtjn\") pod \"ceilometer-0\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.959180 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-scripts\") pod \"ceilometer-0\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.959229 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-config-data\") pod \"ceilometer-0\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.959268 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f41fef51-25cf-4306-87fa-4f203d7d18e9-log-httpd\") pod \"ceilometer-0\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.959291 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.959336 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.959369 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f41fef51-25cf-4306-87fa-4f203d7d18e9-run-httpd\") pod \"ceilometer-0\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.960083 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f41fef51-25cf-4306-87fa-4f203d7d18e9-run-httpd\") pod \"ceilometer-0\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.960375 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f41fef51-25cf-4306-87fa-4f203d7d18e9-log-httpd\") pod \"ceilometer-0\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.964198 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.964913 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-config-data\") pod \"ceilometer-0\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.971553 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-scripts\") pod \"ceilometer-0\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.972519 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " pod="openstack/ceilometer-0" Mar 17 09:32:40 crc kubenswrapper[4813]: I0317 09:32:40.980668 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jtjn\" (UniqueName: \"kubernetes.io/projected/f41fef51-25cf-4306-87fa-4f203d7d18e9-kube-api-access-2jtjn\") pod \"ceilometer-0\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " pod="openstack/ceilometer-0" Mar 17 09:32:41 crc kubenswrapper[4813]: I0317 09:32:41.099316 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:32:41 crc kubenswrapper[4813]: I0317 09:32:41.574144 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:32:42 crc kubenswrapper[4813]: I0317 09:32:42.422264 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f41fef51-25cf-4306-87fa-4f203d7d18e9","Type":"ContainerStarted","Data":"938da892e8016b9e2fd28ea9a2f6d04804e63fdd85adaa92b26e1c84b1116673"} Mar 17 09:32:42 crc kubenswrapper[4813]: I0317 09:32:42.778297 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8efe575c-ec1e-4678-8c52-db4dd272d4d1" path="/var/lib/kubelet/pods/8efe575c-ec1e-4678-8c52-db4dd272d4d1/volumes" Mar 17 09:32:43 crc kubenswrapper[4813]: I0317 09:32:43.432918 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f41fef51-25cf-4306-87fa-4f203d7d18e9","Type":"ContainerStarted","Data":"01eefccb4733f564717eef9cd02109e444797b79107d8c82d5c9cdf662bdbe39"} Mar 17 09:32:43 crc kubenswrapper[4813]: I0317 09:32:43.433221 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f41fef51-25cf-4306-87fa-4f203d7d18e9","Type":"ContainerStarted","Data":"82687c8db83081af109ede7b426e5712fbcae4c4e3a3097ca211e27cbf1fe77e"} Mar 17 09:32:44 crc kubenswrapper[4813]: I0317 09:32:44.114489 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:32:44 crc kubenswrapper[4813]: I0317 09:32:44.114564 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:32:44 crc kubenswrapper[4813]: I0317 09:32:44.451780 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f41fef51-25cf-4306-87fa-4f203d7d18e9","Type":"ContainerStarted","Data":"0533d88c48b4cf268e85d8b71e73c298620f3f3e352445936593f80fb3c9954d"} Mar 17 09:32:44 crc kubenswrapper[4813]: I0317 09:32:44.767496 4813 scope.go:117] "RemoveContainer" containerID="60a4c6e72559b0d6e67762dad3d98f179738de32e3fbf9bfcb1db03da4733ced" Mar 17 09:32:46 crc kubenswrapper[4813]: I0317 09:32:46.472157 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f41fef51-25cf-4306-87fa-4f203d7d18e9","Type":"ContainerStarted","Data":"a6b6a160c064ec98158a0a506fb892a51ee338a0fbf0fe48426cb9499e167a73"} Mar 17 09:32:46 crc kubenswrapper[4813]: I0317 09:32:46.473678 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 17 09:32:46 crc kubenswrapper[4813]: I0317 09:32:46.503041 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.058508218 podStartE2EDuration="6.503020651s" podCreationTimestamp="2026-03-17 09:32:40 +0000 UTC" firstStartedPulling="2026-03-17 09:32:41.573565116 +0000 UTC m=+1383.674368655" lastFinishedPulling="2026-03-17 09:32:46.018077589 +0000 UTC m=+1388.118881088" observedRunningTime="2026-03-17 09:32:46.501679059 +0000 UTC m=+1388.602482578" watchObservedRunningTime="2026-03-17 09:32:46.503020651 +0000 UTC m=+1388.603824150" Mar 17 09:32:49 crc kubenswrapper[4813]: I0317 09:32:49.501704 4813 generic.go:334] "Generic (PLEG): container finished" podID="48793e32-107b-4ba7-b3a6-3e8172306408" containerID="cd051cb78eab1e8bb6bdda3cac209d8aa12d7ecdf22cc87948d4018db9a49829" exitCode=0 Mar 17 09:32:49 crc kubenswrapper[4813]: I0317 09:32:49.501945 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jdtxs" event={"ID":"48793e32-107b-4ba7-b3a6-3e8172306408","Type":"ContainerDied","Data":"cd051cb78eab1e8bb6bdda3cac209d8aa12d7ecdf22cc87948d4018db9a49829"} Mar 17 09:32:50 crc kubenswrapper[4813]: I0317 09:32:50.917077 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jdtxs" Mar 17 09:32:50 crc kubenswrapper[4813]: I0317 09:32:50.953689 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48793e32-107b-4ba7-b3a6-3e8172306408-config-data\") pod \"48793e32-107b-4ba7-b3a6-3e8172306408\" (UID: \"48793e32-107b-4ba7-b3a6-3e8172306408\") " Mar 17 09:32:50 crc kubenswrapper[4813]: I0317 09:32:50.954026 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzgvx\" (UniqueName: \"kubernetes.io/projected/48793e32-107b-4ba7-b3a6-3e8172306408-kube-api-access-nzgvx\") pod \"48793e32-107b-4ba7-b3a6-3e8172306408\" (UID: \"48793e32-107b-4ba7-b3a6-3e8172306408\") " Mar 17 09:32:50 crc kubenswrapper[4813]: I0317 09:32:50.954179 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48793e32-107b-4ba7-b3a6-3e8172306408-combined-ca-bundle\") pod \"48793e32-107b-4ba7-b3a6-3e8172306408\" (UID: \"48793e32-107b-4ba7-b3a6-3e8172306408\") " Mar 17 09:32:50 crc kubenswrapper[4813]: I0317 09:32:50.954243 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48793e32-107b-4ba7-b3a6-3e8172306408-scripts\") pod \"48793e32-107b-4ba7-b3a6-3e8172306408\" (UID: \"48793e32-107b-4ba7-b3a6-3e8172306408\") " Mar 17 09:32:50 crc kubenswrapper[4813]: I0317 09:32:50.968740 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48793e32-107b-4ba7-b3a6-3e8172306408-scripts" (OuterVolumeSpecName: "scripts") pod "48793e32-107b-4ba7-b3a6-3e8172306408" (UID: "48793e32-107b-4ba7-b3a6-3e8172306408"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:50 crc kubenswrapper[4813]: I0317 09:32:50.977892 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48793e32-107b-4ba7-b3a6-3e8172306408-kube-api-access-nzgvx" (OuterVolumeSpecName: "kube-api-access-nzgvx") pod "48793e32-107b-4ba7-b3a6-3e8172306408" (UID: "48793e32-107b-4ba7-b3a6-3e8172306408"). InnerVolumeSpecName "kube-api-access-nzgvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:32:50 crc kubenswrapper[4813]: I0317 09:32:50.989330 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48793e32-107b-4ba7-b3a6-3e8172306408-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "48793e32-107b-4ba7-b3a6-3e8172306408" (UID: "48793e32-107b-4ba7-b3a6-3e8172306408"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.002620 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48793e32-107b-4ba7-b3a6-3e8172306408-config-data" (OuterVolumeSpecName: "config-data") pod "48793e32-107b-4ba7-b3a6-3e8172306408" (UID: "48793e32-107b-4ba7-b3a6-3e8172306408"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.056344 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzgvx\" (UniqueName: \"kubernetes.io/projected/48793e32-107b-4ba7-b3a6-3e8172306408-kube-api-access-nzgvx\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.056382 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/48793e32-107b-4ba7-b3a6-3e8172306408-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.056397 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/48793e32-107b-4ba7-b3a6-3e8172306408-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.056409 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/48793e32-107b-4ba7-b3a6-3e8172306408-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.531308 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-jdtxs" event={"ID":"48793e32-107b-4ba7-b3a6-3e8172306408","Type":"ContainerDied","Data":"21fca64e599b1bac11a68ee1c8f71233c7ddc665382163caf82a92a51bfb5408"} Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.531370 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21fca64e599b1bac11a68ee1c8f71233c7ddc665382163caf82a92a51bfb5408" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.531388 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-jdtxs" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.650553 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 17 09:32:51 crc kubenswrapper[4813]: E0317 09:32:51.651117 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48793e32-107b-4ba7-b3a6-3e8172306408" containerName="nova-cell0-conductor-db-sync" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.651138 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="48793e32-107b-4ba7-b3a6-3e8172306408" containerName="nova-cell0-conductor-db-sync" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.651388 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="48793e32-107b-4ba7-b3a6-3e8172306408" containerName="nova-cell0-conductor-db-sync" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.652141 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.656854 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.657368 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-k8fzw" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.667313 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.770670 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08fc5042-2349-4425-a338-e9f9ab5cc125-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"08fc5042-2349-4425-a338-e9f9ab5cc125\") " pod="openstack/nova-cell0-conductor-0" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.770842 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08fc5042-2349-4425-a338-e9f9ab5cc125-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"08fc5042-2349-4425-a338-e9f9ab5cc125\") " pod="openstack/nova-cell0-conductor-0" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.770879 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kdhs2\" (UniqueName: \"kubernetes.io/projected/08fc5042-2349-4425-a338-e9f9ab5cc125-kube-api-access-kdhs2\") pod \"nova-cell0-conductor-0\" (UID: \"08fc5042-2349-4425-a338-e9f9ab5cc125\") " pod="openstack/nova-cell0-conductor-0" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.873410 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08fc5042-2349-4425-a338-e9f9ab5cc125-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"08fc5042-2349-4425-a338-e9f9ab5cc125\") " pod="openstack/nova-cell0-conductor-0" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.873488 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kdhs2\" (UniqueName: \"kubernetes.io/projected/08fc5042-2349-4425-a338-e9f9ab5cc125-kube-api-access-kdhs2\") pod \"nova-cell0-conductor-0\" (UID: \"08fc5042-2349-4425-a338-e9f9ab5cc125\") " pod="openstack/nova-cell0-conductor-0" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.873821 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08fc5042-2349-4425-a338-e9f9ab5cc125-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"08fc5042-2349-4425-a338-e9f9ab5cc125\") " pod="openstack/nova-cell0-conductor-0" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.878625 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/08fc5042-2349-4425-a338-e9f9ab5cc125-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"08fc5042-2349-4425-a338-e9f9ab5cc125\") " pod="openstack/nova-cell0-conductor-0" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.881505 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/08fc5042-2349-4425-a338-e9f9ab5cc125-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"08fc5042-2349-4425-a338-e9f9ab5cc125\") " pod="openstack/nova-cell0-conductor-0" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.895235 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kdhs2\" (UniqueName: \"kubernetes.io/projected/08fc5042-2349-4425-a338-e9f9ab5cc125-kube-api-access-kdhs2\") pod \"nova-cell0-conductor-0\" (UID: \"08fc5042-2349-4425-a338-e9f9ab5cc125\") " pod="openstack/nova-cell0-conductor-0" Mar 17 09:32:51 crc kubenswrapper[4813]: I0317 09:32:51.972859 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Mar 17 09:32:52 crc kubenswrapper[4813]: I0317 09:32:52.509166 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Mar 17 09:32:52 crc kubenswrapper[4813]: W0317 09:32:52.515876 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod08fc5042_2349_4425_a338_e9f9ab5cc125.slice/crio-ec5d38165719d5ca115aed80f7d7da571f32fcb5cef1a143a1740206c0edab89 WatchSource:0}: Error finding container ec5d38165719d5ca115aed80f7d7da571f32fcb5cef1a143a1740206c0edab89: Status 404 returned error can't find the container with id ec5d38165719d5ca115aed80f7d7da571f32fcb5cef1a143a1740206c0edab89 Mar 17 09:32:52 crc kubenswrapper[4813]: I0317 09:32:52.546198 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"08fc5042-2349-4425-a338-e9f9ab5cc125","Type":"ContainerStarted","Data":"ec5d38165719d5ca115aed80f7d7da571f32fcb5cef1a143a1740206c0edab89"} Mar 17 09:32:53 crc kubenswrapper[4813]: I0317 09:32:53.562356 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"08fc5042-2349-4425-a338-e9f9ab5cc125","Type":"ContainerStarted","Data":"62070b615e96598e48d2743ed6a53a5a2a2f0bc5080ce270f9ad936457757f73"} Mar 17 09:32:53 crc kubenswrapper[4813]: I0317 09:32:53.562765 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Mar 17 09:32:53 crc kubenswrapper[4813]: I0317 09:32:53.605234 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.60520702 podStartE2EDuration="2.60520702s" podCreationTimestamp="2026-03-17 09:32:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:32:53.581141314 +0000 UTC m=+1395.681944853" watchObservedRunningTime="2026-03-17 09:32:53.60520702 +0000 UTC m=+1395.706010559" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.017306 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.520814 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-2j8vj"] Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.522269 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2j8vj" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.524437 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.527452 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.539466 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-2j8vj"] Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.598493 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n5dw\" (UniqueName: \"kubernetes.io/projected/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-kube-api-access-6n5dw\") pod \"nova-cell0-cell-mapping-2j8vj\" (UID: \"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79\") " pod="openstack/nova-cell0-cell-mapping-2j8vj" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.598756 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-config-data\") pod \"nova-cell0-cell-mapping-2j8vj\" (UID: \"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79\") " pod="openstack/nova-cell0-cell-mapping-2j8vj" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.598853 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-scripts\") pod \"nova-cell0-cell-mapping-2j8vj\" (UID: \"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79\") " pod="openstack/nova-cell0-cell-mapping-2j8vj" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.598882 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-2j8vj\" (UID: \"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79\") " pod="openstack/nova-cell0-cell-mapping-2j8vj" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.697092 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.699072 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.700723 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n5dw\" (UniqueName: \"kubernetes.io/projected/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-kube-api-access-6n5dw\") pod \"nova-cell0-cell-mapping-2j8vj\" (UID: \"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79\") " pod="openstack/nova-cell0-cell-mapping-2j8vj" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.700837 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-config-data\") pod \"nova-cell0-cell-mapping-2j8vj\" (UID: \"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79\") " pod="openstack/nova-cell0-cell-mapping-2j8vj" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.700895 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-scripts\") pod \"nova-cell0-cell-mapping-2j8vj\" (UID: \"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79\") " pod="openstack/nova-cell0-cell-mapping-2j8vj" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.700914 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-2j8vj\" (UID: \"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79\") " pod="openstack/nova-cell0-cell-mapping-2j8vj" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.722836 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-scripts\") pod \"nova-cell0-cell-mapping-2j8vj\" (UID: \"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79\") " pod="openstack/nova-cell0-cell-mapping-2j8vj" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.731940 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.734055 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-2j8vj\" (UID: \"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79\") " pod="openstack/nova-cell0-cell-mapping-2j8vj" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.734160 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-config-data\") pod \"nova-cell0-cell-mapping-2j8vj\" (UID: \"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79\") " pod="openstack/nova-cell0-cell-mapping-2j8vj" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.738171 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n5dw\" (UniqueName: \"kubernetes.io/projected/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-kube-api-access-6n5dw\") pod \"nova-cell0-cell-mapping-2j8vj\" (UID: \"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79\") " pod="openstack/nova-cell0-cell-mapping-2j8vj" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.774147 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.783713 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.785338 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.789851 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.802158 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e\") " pod="openstack/nova-metadata-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.802211 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d318c13f-5f03-40af-86a1-4fd58b6f6b70-logs\") pod \"nova-api-0\" (UID: \"d318c13f-5f03-40af-86a1-4fd58b6f6b70\") " pod="openstack/nova-api-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.802376 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-config-data\") pod \"nova-metadata-0\" (UID: \"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e\") " pod="openstack/nova-metadata-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.802438 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pd78\" (UniqueName: \"kubernetes.io/projected/d318c13f-5f03-40af-86a1-4fd58b6f6b70-kube-api-access-8pd78\") pod \"nova-api-0\" (UID: \"d318c13f-5f03-40af-86a1-4fd58b6f6b70\") " pod="openstack/nova-api-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.802475 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d318c13f-5f03-40af-86a1-4fd58b6f6b70-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d318c13f-5f03-40af-86a1-4fd58b6f6b70\") " pod="openstack/nova-api-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.802501 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d318c13f-5f03-40af-86a1-4fd58b6f6b70-config-data\") pod \"nova-api-0\" (UID: \"d318c13f-5f03-40af-86a1-4fd58b6f6b70\") " pod="openstack/nova-api-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.802558 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lj85\" (UniqueName: \"kubernetes.io/projected/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-kube-api-access-6lj85\") pod \"nova-metadata-0\" (UID: \"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e\") " pod="openstack/nova-metadata-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.802668 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-logs\") pod \"nova-metadata-0\" (UID: \"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e\") " pod="openstack/nova-metadata-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.834239 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.857478 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2j8vj" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.896650 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.897810 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.900232 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.904660 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pd78\" (UniqueName: \"kubernetes.io/projected/d318c13f-5f03-40af-86a1-4fd58b6f6b70-kube-api-access-8pd78\") pod \"nova-api-0\" (UID: \"d318c13f-5f03-40af-86a1-4fd58b6f6b70\") " pod="openstack/nova-api-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.904698 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baca76f9-d625-4532-9315-8be77971153b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"baca76f9-d625-4532-9315-8be77971153b\") " pod="openstack/nova-scheduler-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.904722 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d318c13f-5f03-40af-86a1-4fd58b6f6b70-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d318c13f-5f03-40af-86a1-4fd58b6f6b70\") " pod="openstack/nova-api-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.904737 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d318c13f-5f03-40af-86a1-4fd58b6f6b70-config-data\") pod \"nova-api-0\" (UID: \"d318c13f-5f03-40af-86a1-4fd58b6f6b70\") " pod="openstack/nova-api-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.904772 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lj85\" (UniqueName: \"kubernetes.io/projected/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-kube-api-access-6lj85\") pod \"nova-metadata-0\" (UID: \"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e\") " pod="openstack/nova-metadata-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.904814 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljfmb\" (UniqueName: \"kubernetes.io/projected/baca76f9-d625-4532-9315-8be77971153b-kube-api-access-ljfmb\") pod \"nova-scheduler-0\" (UID: \"baca76f9-d625-4532-9315-8be77971153b\") " pod="openstack/nova-scheduler-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.904838 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-logs\") pod \"nova-metadata-0\" (UID: \"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e\") " pod="openstack/nova-metadata-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.904859 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e\") " pod="openstack/nova-metadata-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.904874 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d318c13f-5f03-40af-86a1-4fd58b6f6b70-logs\") pod \"nova-api-0\" (UID: \"d318c13f-5f03-40af-86a1-4fd58b6f6b70\") " pod="openstack/nova-api-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.904941 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-config-data\") pod \"nova-metadata-0\" (UID: \"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e\") " pod="openstack/nova-metadata-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.904960 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baca76f9-d625-4532-9315-8be77971153b-config-data\") pod \"nova-scheduler-0\" (UID: \"baca76f9-d625-4532-9315-8be77971153b\") " pod="openstack/nova-scheduler-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.907003 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-logs\") pod \"nova-metadata-0\" (UID: \"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e\") " pod="openstack/nova-metadata-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.911077 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d318c13f-5f03-40af-86a1-4fd58b6f6b70-logs\") pod \"nova-api-0\" (UID: \"d318c13f-5f03-40af-86a1-4fd58b6f6b70\") " pod="openstack/nova-api-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.920137 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-k9crm"] Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.925855 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.928237 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d318c13f-5f03-40af-86a1-4fd58b6f6b70-config-data\") pod \"nova-api-0\" (UID: \"d318c13f-5f03-40af-86a1-4fd58b6f6b70\") " pod="openstack/nova-api-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.934189 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d318c13f-5f03-40af-86a1-4fd58b6f6b70-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"d318c13f-5f03-40af-86a1-4fd58b6f6b70\") " pod="openstack/nova-api-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.936831 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pd78\" (UniqueName: \"kubernetes.io/projected/d318c13f-5f03-40af-86a1-4fd58b6f6b70-kube-api-access-8pd78\") pod \"nova-api-0\" (UID: \"d318c13f-5f03-40af-86a1-4fd58b6f6b70\") " pod="openstack/nova-api-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.942346 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e\") " pod="openstack/nova-metadata-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.942848 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-config-data\") pod \"nova-metadata-0\" (UID: \"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e\") " pod="openstack/nova-metadata-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.944230 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lj85\" (UniqueName: \"kubernetes.io/projected/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-kube-api-access-6lj85\") pod \"nova-metadata-0\" (UID: \"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e\") " pod="openstack/nova-metadata-0" Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.971466 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 09:33:02 crc kubenswrapper[4813]: I0317 09:33:02.989451 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-k9crm"] Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.006271 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baca76f9-d625-4532-9315-8be77971153b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"baca76f9-d625-4532-9315-8be77971153b\") " pod="openstack/nova-scheduler-0" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.006378 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-k9crm\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.006418 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljfmb\" (UniqueName: \"kubernetes.io/projected/baca76f9-d625-4532-9315-8be77971153b-kube-api-access-ljfmb\") pod \"nova-scheduler-0\" (UID: \"baca76f9-d625-4532-9315-8be77971153b\") " pod="openstack/nova-scheduler-0" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.006447 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-k9crm\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.006500 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5xx6\" (UniqueName: \"kubernetes.io/projected/7666a75c-63d7-4a2f-9280-4ec426a02358-kube-api-access-v5xx6\") pod \"dnsmasq-dns-845d6d6f59-k9crm\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.006550 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-k9crm\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.006571 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-config\") pod \"dnsmasq-dns-845d6d6f59-k9crm\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.006626 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baca76f9-d625-4532-9315-8be77971153b-config-data\") pod \"nova-scheduler-0\" (UID: \"baca76f9-d625-4532-9315-8be77971153b\") " pod="openstack/nova-scheduler-0" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.006668 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-k9crm\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.023793 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baca76f9-d625-4532-9315-8be77971153b-config-data\") pod \"nova-scheduler-0\" (UID: \"baca76f9-d625-4532-9315-8be77971153b\") " pod="openstack/nova-scheduler-0" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.024343 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baca76f9-d625-4532-9315-8be77971153b-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"baca76f9-d625-4532-9315-8be77971153b\") " pod="openstack/nova-scheduler-0" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.032376 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.034029 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.037419 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.041304 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.043576 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljfmb\" (UniqueName: \"kubernetes.io/projected/baca76f9-d625-4532-9315-8be77971153b-kube-api-access-ljfmb\") pod \"nova-scheduler-0\" (UID: \"baca76f9-d625-4532-9315-8be77971153b\") " pod="openstack/nova-scheduler-0" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.055056 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.115702 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5xx6\" (UniqueName: \"kubernetes.io/projected/7666a75c-63d7-4a2f-9280-4ec426a02358-kube-api-access-v5xx6\") pod \"dnsmasq-dns-845d6d6f59-k9crm\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.115973 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-k9crm\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.115992 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-config\") pod \"dnsmasq-dns-845d6d6f59-k9crm\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.116036 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsj52\" (UniqueName: \"kubernetes.io/projected/845fdf7b-a3b0-487b-9bcd-efcc3b2b8513-kube-api-access-lsj52\") pod \"nova-cell1-novncproxy-0\" (UID: \"845fdf7b-a3b0-487b-9bcd-efcc3b2b8513\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.116070 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-k9crm\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.116111 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/845fdf7b-a3b0-487b-9bcd-efcc3b2b8513-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"845fdf7b-a3b0-487b-9bcd-efcc3b2b8513\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.116141 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/845fdf7b-a3b0-487b-9bcd-efcc3b2b8513-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"845fdf7b-a3b0-487b-9bcd-efcc3b2b8513\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.116187 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-k9crm\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.116228 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-k9crm\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.117036 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-ovsdbserver-sb\") pod \"dnsmasq-dns-845d6d6f59-k9crm\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.117558 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-ovsdbserver-nb\") pod \"dnsmasq-dns-845d6d6f59-k9crm\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.117876 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-dns-svc\") pod \"dnsmasq-dns-845d6d6f59-k9crm\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.118374 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-config\") pod \"dnsmasq-dns-845d6d6f59-k9crm\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.118959 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-dns-swift-storage-0\") pod \"dnsmasq-dns-845d6d6f59-k9crm\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.137684 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.146021 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5xx6\" (UniqueName: \"kubernetes.io/projected/7666a75c-63d7-4a2f-9280-4ec426a02358-kube-api-access-v5xx6\") pod \"dnsmasq-dns-845d6d6f59-k9crm\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.157022 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.227284 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsj52\" (UniqueName: \"kubernetes.io/projected/845fdf7b-a3b0-487b-9bcd-efcc3b2b8513-kube-api-access-lsj52\") pod \"nova-cell1-novncproxy-0\" (UID: \"845fdf7b-a3b0-487b-9bcd-efcc3b2b8513\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.227362 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/845fdf7b-a3b0-487b-9bcd-efcc3b2b8513-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"845fdf7b-a3b0-487b-9bcd-efcc3b2b8513\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.227394 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/845fdf7b-a3b0-487b-9bcd-efcc3b2b8513-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"845fdf7b-a3b0-487b-9bcd-efcc3b2b8513\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.231067 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/845fdf7b-a3b0-487b-9bcd-efcc3b2b8513-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"845fdf7b-a3b0-487b-9bcd-efcc3b2b8513\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.233051 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/845fdf7b-a3b0-487b-9bcd-efcc3b2b8513-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"845fdf7b-a3b0-487b-9bcd-efcc3b2b8513\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.249513 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsj52\" (UniqueName: \"kubernetes.io/projected/845fdf7b-a3b0-487b-9bcd-efcc3b2b8513-kube-api-access-lsj52\") pod \"nova-cell1-novncproxy-0\" (UID: \"845fdf7b-a3b0-487b-9bcd-efcc3b2b8513\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.366001 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.374357 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.439411 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vqcqs"] Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.440642 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vqcqs" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.443854 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.444059 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.450884 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vqcqs"] Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.511304 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-2j8vj"] Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.532120 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/500133a3-3767-4bd8-8011-066d1b77ddca-scripts\") pod \"nova-cell1-conductor-db-sync-vqcqs\" (UID: \"500133a3-3767-4bd8-8011-066d1b77ddca\") " pod="openstack/nova-cell1-conductor-db-sync-vqcqs" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.532192 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/500133a3-3767-4bd8-8011-066d1b77ddca-config-data\") pod \"nova-cell1-conductor-db-sync-vqcqs\" (UID: \"500133a3-3767-4bd8-8011-066d1b77ddca\") " pod="openstack/nova-cell1-conductor-db-sync-vqcqs" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.532249 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/500133a3-3767-4bd8-8011-066d1b77ddca-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vqcqs\" (UID: \"500133a3-3767-4bd8-8011-066d1b77ddca\") " pod="openstack/nova-cell1-conductor-db-sync-vqcqs" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.532308 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6v7d\" (UniqueName: \"kubernetes.io/projected/500133a3-3767-4bd8-8011-066d1b77ddca-kube-api-access-b6v7d\") pod \"nova-cell1-conductor-db-sync-vqcqs\" (UID: \"500133a3-3767-4bd8-8011-066d1b77ddca\") " pod="openstack/nova-cell1-conductor-db-sync-vqcqs" Mar 17 09:33:03 crc kubenswrapper[4813]: W0317 09:33:03.539099 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd08fdcd8_7f6c_468e_bc05_5f0c8a4add79.slice/crio-6094dff23473783db9834ac407f17f80a71a7c2141c3fe7d0ae6ed4f4c911da8 WatchSource:0}: Error finding container 6094dff23473783db9834ac407f17f80a71a7c2141c3fe7d0ae6ed4f4c911da8: Status 404 returned error can't find the container with id 6094dff23473783db9834ac407f17f80a71a7c2141c3fe7d0ae6ed4f4c911da8 Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.625838 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.633917 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/500133a3-3767-4bd8-8011-066d1b77ddca-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vqcqs\" (UID: \"500133a3-3767-4bd8-8011-066d1b77ddca\") " pod="openstack/nova-cell1-conductor-db-sync-vqcqs" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.634009 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6v7d\" (UniqueName: \"kubernetes.io/projected/500133a3-3767-4bd8-8011-066d1b77ddca-kube-api-access-b6v7d\") pod \"nova-cell1-conductor-db-sync-vqcqs\" (UID: \"500133a3-3767-4bd8-8011-066d1b77ddca\") " pod="openstack/nova-cell1-conductor-db-sync-vqcqs" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.634056 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/500133a3-3767-4bd8-8011-066d1b77ddca-scripts\") pod \"nova-cell1-conductor-db-sync-vqcqs\" (UID: \"500133a3-3767-4bd8-8011-066d1b77ddca\") " pod="openstack/nova-cell1-conductor-db-sync-vqcqs" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.634099 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/500133a3-3767-4bd8-8011-066d1b77ddca-config-data\") pod \"nova-cell1-conductor-db-sync-vqcqs\" (UID: \"500133a3-3767-4bd8-8011-066d1b77ddca\") " pod="openstack/nova-cell1-conductor-db-sync-vqcqs" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.638675 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/500133a3-3767-4bd8-8011-066d1b77ddca-scripts\") pod \"nova-cell1-conductor-db-sync-vqcqs\" (UID: \"500133a3-3767-4bd8-8011-066d1b77ddca\") " pod="openstack/nova-cell1-conductor-db-sync-vqcqs" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.638834 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/500133a3-3767-4bd8-8011-066d1b77ddca-config-data\") pod \"nova-cell1-conductor-db-sync-vqcqs\" (UID: \"500133a3-3767-4bd8-8011-066d1b77ddca\") " pod="openstack/nova-cell1-conductor-db-sync-vqcqs" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.644401 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.645684 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/500133a3-3767-4bd8-8011-066d1b77ddca-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-vqcqs\" (UID: \"500133a3-3767-4bd8-8011-066d1b77ddca\") " pod="openstack/nova-cell1-conductor-db-sync-vqcqs" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.658081 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6v7d\" (UniqueName: \"kubernetes.io/projected/500133a3-3767-4bd8-8011-066d1b77ddca-kube-api-access-b6v7d\") pod \"nova-cell1-conductor-db-sync-vqcqs\" (UID: \"500133a3-3767-4bd8-8011-066d1b77ddca\") " pod="openstack/nova-cell1-conductor-db-sync-vqcqs" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.688466 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d318c13f-5f03-40af-86a1-4fd58b6f6b70","Type":"ContainerStarted","Data":"412a220eea0d0cfb004a96cdb0560254828abb5865076380779b6ca448167f7a"} Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.689679 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"baca76f9-d625-4532-9315-8be77971153b","Type":"ContainerStarted","Data":"3b878f5897f8fb9a0a746d3e5228b91849e593d27460757caf83ab7c1fe680f9"} Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.690534 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2j8vj" event={"ID":"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79","Type":"ContainerStarted","Data":"6094dff23473783db9834ac407f17f80a71a7c2141c3fe7d0ae6ed4f4c911da8"} Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.768580 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vqcqs" Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.815812 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 09:33:03 crc kubenswrapper[4813]: W0317 09:33:03.829498 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5b6fe5fb_6ea8_4b5c_96d1_68148dbec03e.slice/crio-faaa4aed572ad4dab826885fef1f11c5f04748dab1daca2adb327dc69f0e8914 WatchSource:0}: Error finding container faaa4aed572ad4dab826885fef1f11c5f04748dab1daca2adb327dc69f0e8914: Status 404 returned error can't find the container with id faaa4aed572ad4dab826885fef1f11c5f04748dab1daca2adb327dc69f0e8914 Mar 17 09:33:03 crc kubenswrapper[4813]: I0317 09:33:03.985048 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-k9crm"] Mar 17 09:33:04 crc kubenswrapper[4813]: I0317 09:33:04.038160 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 17 09:33:04 crc kubenswrapper[4813]: I0317 09:33:04.261324 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vqcqs"] Mar 17 09:33:04 crc kubenswrapper[4813]: W0317 09:33:04.313962 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod500133a3_3767_4bd8_8011_066d1b77ddca.slice/crio-7324d982259930e6c1d95232e3056e4ef1b9d2d2d6a9f01a3b30b57066151a32 WatchSource:0}: Error finding container 7324d982259930e6c1d95232e3056e4ef1b9d2d2d6a9f01a3b30b57066151a32: Status 404 returned error can't find the container with id 7324d982259930e6c1d95232e3056e4ef1b9d2d2d6a9f01a3b30b57066151a32 Mar 17 09:33:04 crc kubenswrapper[4813]: I0317 09:33:04.726895 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vqcqs" event={"ID":"500133a3-3767-4bd8-8011-066d1b77ddca","Type":"ContainerStarted","Data":"3671bbe1c1f3b2c6bde24a80e90ff0a32ab448a617d0b0d5858aeb5ce6fe5c63"} Mar 17 09:33:04 crc kubenswrapper[4813]: I0317 09:33:04.727363 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vqcqs" event={"ID":"500133a3-3767-4bd8-8011-066d1b77ddca","Type":"ContainerStarted","Data":"7324d982259930e6c1d95232e3056e4ef1b9d2d2d6a9f01a3b30b57066151a32"} Mar 17 09:33:04 crc kubenswrapper[4813]: I0317 09:33:04.730460 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2j8vj" event={"ID":"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79","Type":"ContainerStarted","Data":"ebdf2c0cb40a2d460d25104c47216d741f61c3ee45bfae8c282b3ef6465331d1"} Mar 17 09:33:04 crc kubenswrapper[4813]: I0317 09:33:04.732957 4813 generic.go:334] "Generic (PLEG): container finished" podID="7666a75c-63d7-4a2f-9280-4ec426a02358" containerID="d5b1a865da51567bf15798bce886e42ac5f2b60764cec2deb2ffced06ca72f62" exitCode=0 Mar 17 09:33:04 crc kubenswrapper[4813]: I0317 09:33:04.742994 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" event={"ID":"7666a75c-63d7-4a2f-9280-4ec426a02358","Type":"ContainerDied","Data":"d5b1a865da51567bf15798bce886e42ac5f2b60764cec2deb2ffced06ca72f62"} Mar 17 09:33:04 crc kubenswrapper[4813]: I0317 09:33:04.743033 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" event={"ID":"7666a75c-63d7-4a2f-9280-4ec426a02358","Type":"ContainerStarted","Data":"59725a6479c946c1029cbff3f9dae15747e3bee03b822295a7d17081b71fda3d"} Mar 17 09:33:04 crc kubenswrapper[4813]: I0317 09:33:04.743045 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"845fdf7b-a3b0-487b-9bcd-efcc3b2b8513","Type":"ContainerStarted","Data":"e7013bd9fd2db2fb53e84999c9ed74477919c76cacc8da0eec310d8b91807f8b"} Mar 17 09:33:04 crc kubenswrapper[4813]: I0317 09:33:04.743121 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e","Type":"ContainerStarted","Data":"faaa4aed572ad4dab826885fef1f11c5f04748dab1daca2adb327dc69f0e8914"} Mar 17 09:33:04 crc kubenswrapper[4813]: I0317 09:33:04.756504 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-vqcqs" podStartSLOduration=1.7564900159999999 podStartE2EDuration="1.756490016s" podCreationTimestamp="2026-03-17 09:33:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:33:04.74896295 +0000 UTC m=+1406.849766449" watchObservedRunningTime="2026-03-17 09:33:04.756490016 +0000 UTC m=+1406.857293515" Mar 17 09:33:04 crc kubenswrapper[4813]: I0317 09:33:04.770927 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-2j8vj" podStartSLOduration=2.77090819 podStartE2EDuration="2.77090819s" podCreationTimestamp="2026-03-17 09:33:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:33:04.762743953 +0000 UTC m=+1406.863547452" watchObservedRunningTime="2026-03-17 09:33:04.77090819 +0000 UTC m=+1406.871711689" Mar 17 09:33:06 crc kubenswrapper[4813]: I0317 09:33:06.459502 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 09:33:06 crc kubenswrapper[4813]: I0317 09:33:06.469971 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 17 09:33:07 crc kubenswrapper[4813]: I0317 09:33:07.775089 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"845fdf7b-a3b0-487b-9bcd-efcc3b2b8513","Type":"ContainerStarted","Data":"987f42380506732cb7bccd1f21a56d81e217d17ebf5e5f8bb56bd3397368fdac"} Mar 17 09:33:07 crc kubenswrapper[4813]: I0317 09:33:07.775235 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="845fdf7b-a3b0-487b-9bcd-efcc3b2b8513" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://987f42380506732cb7bccd1f21a56d81e217d17ebf5e5f8bb56bd3397368fdac" gracePeriod=30 Mar 17 09:33:07 crc kubenswrapper[4813]: I0317 09:33:07.777862 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e","Type":"ContainerStarted","Data":"73e9e45ab00fd186ef839c6120ddef5b2262c385e278391807dc7a69b25374c7"} Mar 17 09:33:07 crc kubenswrapper[4813]: I0317 09:33:07.777927 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e","Type":"ContainerStarted","Data":"0c8fe55a52148629e20a9e9e5dbdf2a2531712a5a85e9fb82c03955309af3b4e"} Mar 17 09:33:07 crc kubenswrapper[4813]: I0317 09:33:07.778083 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e" containerName="nova-metadata-log" containerID="cri-o://0c8fe55a52148629e20a9e9e5dbdf2a2531712a5a85e9fb82c03955309af3b4e" gracePeriod=30 Mar 17 09:33:07 crc kubenswrapper[4813]: I0317 09:33:07.778231 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e" containerName="nova-metadata-metadata" containerID="cri-o://73e9e45ab00fd186ef839c6120ddef5b2262c385e278391807dc7a69b25374c7" gracePeriod=30 Mar 17 09:33:07 crc kubenswrapper[4813]: I0317 09:33:07.790166 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d318c13f-5f03-40af-86a1-4fd58b6f6b70","Type":"ContainerStarted","Data":"997adf8213b3dd1c07c11b89cc0069428f4f4ee873ac17011cab5fe52d0b58a7"} Mar 17 09:33:07 crc kubenswrapper[4813]: I0317 09:33:07.790270 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d318c13f-5f03-40af-86a1-4fd58b6f6b70","Type":"ContainerStarted","Data":"96b00c2c3649f24a02186e3513367217ea4a0c7995fc302aaf961370b6f222d6"} Mar 17 09:33:07 crc kubenswrapper[4813]: I0317 09:33:07.792237 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"baca76f9-d625-4532-9315-8be77971153b","Type":"ContainerStarted","Data":"b7b096e17f5a01a3456741d7e0f29d468e7898533e0b1e825c1bbf1d682b64ed"} Mar 17 09:33:07 crc kubenswrapper[4813]: I0317 09:33:07.794911 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" event={"ID":"7666a75c-63d7-4a2f-9280-4ec426a02358","Type":"ContainerStarted","Data":"ae597b85c32ec4a1fd3b033d11a1426d69c4b8aa2f4964db75b4e7734c51fac1"} Mar 17 09:33:07 crc kubenswrapper[4813]: I0317 09:33:07.796017 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:07 crc kubenswrapper[4813]: I0317 09:33:07.857526 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" podStartSLOduration=5.857501735 podStartE2EDuration="5.857501735s" podCreationTimestamp="2026-03-17 09:33:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:33:07.851055813 +0000 UTC m=+1409.951859322" watchObservedRunningTime="2026-03-17 09:33:07.857501735 +0000 UTC m=+1409.958305244" Mar 17 09:33:07 crc kubenswrapper[4813]: I0317 09:33:07.858934 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.120158234 podStartE2EDuration="5.858924591s" podCreationTimestamp="2026-03-17 09:33:02 +0000 UTC" firstStartedPulling="2026-03-17 09:33:04.023057569 +0000 UTC m=+1406.123861068" lastFinishedPulling="2026-03-17 09:33:06.761823916 +0000 UTC m=+1408.862627425" observedRunningTime="2026-03-17 09:33:07.815656319 +0000 UTC m=+1409.916459828" watchObservedRunningTime="2026-03-17 09:33:07.858924591 +0000 UTC m=+1409.959728090" Mar 17 09:33:07 crc kubenswrapper[4813]: I0317 09:33:07.877181 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.765814759 podStartE2EDuration="5.877164004s" podCreationTimestamp="2026-03-17 09:33:02 +0000 UTC" firstStartedPulling="2026-03-17 09:33:03.649778959 +0000 UTC m=+1405.750582458" lastFinishedPulling="2026-03-17 09:33:06.761128204 +0000 UTC m=+1408.861931703" observedRunningTime="2026-03-17 09:33:07.867865702 +0000 UTC m=+1409.968669201" watchObservedRunningTime="2026-03-17 09:33:07.877164004 +0000 UTC m=+1409.977967503" Mar 17 09:33:07 crc kubenswrapper[4813]: I0317 09:33:07.901399 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.790370672 podStartE2EDuration="5.901359325s" podCreationTimestamp="2026-03-17 09:33:02 +0000 UTC" firstStartedPulling="2026-03-17 09:33:03.65012333 +0000 UTC m=+1405.750926829" lastFinishedPulling="2026-03-17 09:33:06.761111983 +0000 UTC m=+1408.861915482" observedRunningTime="2026-03-17 09:33:07.896305717 +0000 UTC m=+1409.997109216" watchObservedRunningTime="2026-03-17 09:33:07.901359325 +0000 UTC m=+1410.002162824" Mar 17 09:33:07 crc kubenswrapper[4813]: I0317 09:33:07.922662 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.99254766 podStartE2EDuration="5.922640764s" podCreationTimestamp="2026-03-17 09:33:02 +0000 UTC" firstStartedPulling="2026-03-17 09:33:03.841776888 +0000 UTC m=+1405.942580387" lastFinishedPulling="2026-03-17 09:33:06.771869972 +0000 UTC m=+1408.872673491" observedRunningTime="2026-03-17 09:33:07.914111526 +0000 UTC m=+1410.014915025" watchObservedRunningTime="2026-03-17 09:33:07.922640764 +0000 UTC m=+1410.023444273" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.055908 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.374964 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.687244 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.763848 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-combined-ca-bundle\") pod \"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e\" (UID: \"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e\") " Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.763934 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lj85\" (UniqueName: \"kubernetes.io/projected/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-kube-api-access-6lj85\") pod \"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e\" (UID: \"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e\") " Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.763993 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-config-data\") pod \"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e\" (UID: \"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e\") " Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.764207 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-logs\") pod \"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e\" (UID: \"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e\") " Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.764985 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-logs" (OuterVolumeSpecName: "logs") pod "5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e" (UID: "5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.774608 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-kube-api-access-6lj85" (OuterVolumeSpecName: "kube-api-access-6lj85") pod "5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e" (UID: "5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e"). InnerVolumeSpecName "kube-api-access-6lj85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.806917 4813 generic.go:334] "Generic (PLEG): container finished" podID="5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e" containerID="73e9e45ab00fd186ef839c6120ddef5b2262c385e278391807dc7a69b25374c7" exitCode=0 Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.806953 4813 generic.go:334] "Generic (PLEG): container finished" podID="5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e" containerID="0c8fe55a52148629e20a9e9e5dbdf2a2531712a5a85e9fb82c03955309af3b4e" exitCode=143 Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.807273 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e","Type":"ContainerDied","Data":"73e9e45ab00fd186ef839c6120ddef5b2262c385e278391807dc7a69b25374c7"} Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.807305 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e","Type":"ContainerDied","Data":"0c8fe55a52148629e20a9e9e5dbdf2a2531712a5a85e9fb82c03955309af3b4e"} Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.807318 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e","Type":"ContainerDied","Data":"faaa4aed572ad4dab826885fef1f11c5f04748dab1daca2adb327dc69f0e8914"} Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.807336 4813 scope.go:117] "RemoveContainer" containerID="73e9e45ab00fd186ef839c6120ddef5b2262c385e278391807dc7a69b25374c7" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.807485 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.808552 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-config-data" (OuterVolumeSpecName: "config-data") pod "5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e" (UID: "5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.815816 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e" (UID: "5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.867463 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-logs\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.867507 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.867525 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lj85\" (UniqueName: \"kubernetes.io/projected/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-kube-api-access-6lj85\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.867537 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.903258 4813 scope.go:117] "RemoveContainer" containerID="0c8fe55a52148629e20a9e9e5dbdf2a2531712a5a85e9fb82c03955309af3b4e" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.921908 4813 scope.go:117] "RemoveContainer" containerID="73e9e45ab00fd186ef839c6120ddef5b2262c385e278391807dc7a69b25374c7" Mar 17 09:33:08 crc kubenswrapper[4813]: E0317 09:33:08.922371 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73e9e45ab00fd186ef839c6120ddef5b2262c385e278391807dc7a69b25374c7\": container with ID starting with 73e9e45ab00fd186ef839c6120ddef5b2262c385e278391807dc7a69b25374c7 not found: ID does not exist" containerID="73e9e45ab00fd186ef839c6120ddef5b2262c385e278391807dc7a69b25374c7" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.922402 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73e9e45ab00fd186ef839c6120ddef5b2262c385e278391807dc7a69b25374c7"} err="failed to get container status \"73e9e45ab00fd186ef839c6120ddef5b2262c385e278391807dc7a69b25374c7\": rpc error: code = NotFound desc = could not find container \"73e9e45ab00fd186ef839c6120ddef5b2262c385e278391807dc7a69b25374c7\": container with ID starting with 73e9e45ab00fd186ef839c6120ddef5b2262c385e278391807dc7a69b25374c7 not found: ID does not exist" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.922424 4813 scope.go:117] "RemoveContainer" containerID="0c8fe55a52148629e20a9e9e5dbdf2a2531712a5a85e9fb82c03955309af3b4e" Mar 17 09:33:08 crc kubenswrapper[4813]: E0317 09:33:08.923081 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0c8fe55a52148629e20a9e9e5dbdf2a2531712a5a85e9fb82c03955309af3b4e\": container with ID starting with 0c8fe55a52148629e20a9e9e5dbdf2a2531712a5a85e9fb82c03955309af3b4e not found: ID does not exist" containerID="0c8fe55a52148629e20a9e9e5dbdf2a2531712a5a85e9fb82c03955309af3b4e" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.923106 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c8fe55a52148629e20a9e9e5dbdf2a2531712a5a85e9fb82c03955309af3b4e"} err="failed to get container status \"0c8fe55a52148629e20a9e9e5dbdf2a2531712a5a85e9fb82c03955309af3b4e\": rpc error: code = NotFound desc = could not find container \"0c8fe55a52148629e20a9e9e5dbdf2a2531712a5a85e9fb82c03955309af3b4e\": container with ID starting with 0c8fe55a52148629e20a9e9e5dbdf2a2531712a5a85e9fb82c03955309af3b4e not found: ID does not exist" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.923119 4813 scope.go:117] "RemoveContainer" containerID="73e9e45ab00fd186ef839c6120ddef5b2262c385e278391807dc7a69b25374c7" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.923440 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73e9e45ab00fd186ef839c6120ddef5b2262c385e278391807dc7a69b25374c7"} err="failed to get container status \"73e9e45ab00fd186ef839c6120ddef5b2262c385e278391807dc7a69b25374c7\": rpc error: code = NotFound desc = could not find container \"73e9e45ab00fd186ef839c6120ddef5b2262c385e278391807dc7a69b25374c7\": container with ID starting with 73e9e45ab00fd186ef839c6120ddef5b2262c385e278391807dc7a69b25374c7 not found: ID does not exist" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.923457 4813 scope.go:117] "RemoveContainer" containerID="0c8fe55a52148629e20a9e9e5dbdf2a2531712a5a85e9fb82c03955309af3b4e" Mar 17 09:33:08 crc kubenswrapper[4813]: I0317 09:33:08.923993 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0c8fe55a52148629e20a9e9e5dbdf2a2531712a5a85e9fb82c03955309af3b4e"} err="failed to get container status \"0c8fe55a52148629e20a9e9e5dbdf2a2531712a5a85e9fb82c03955309af3b4e\": rpc error: code = NotFound desc = could not find container \"0c8fe55a52148629e20a9e9e5dbdf2a2531712a5a85e9fb82c03955309af3b4e\": container with ID starting with 0c8fe55a52148629e20a9e9e5dbdf2a2531712a5a85e9fb82c03955309af3b4e not found: ID does not exist" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.146975 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.156162 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.190200 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 17 09:33:09 crc kubenswrapper[4813]: E0317 09:33:09.190944 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e" containerName="nova-metadata-log" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.190979 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e" containerName="nova-metadata-log" Mar 17 09:33:09 crc kubenswrapper[4813]: E0317 09:33:09.191002 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e" containerName="nova-metadata-metadata" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.191017 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e" containerName="nova-metadata-metadata" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.191367 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e" containerName="nova-metadata-metadata" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.191392 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e" containerName="nova-metadata-log" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.193126 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.196752 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.197076 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.210435 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.276300 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk8f4\" (UniqueName: \"kubernetes.io/projected/d59afa61-6c33-432d-822a-da4a1c6519bf-kube-api-access-pk8f4\") pod \"nova-metadata-0\" (UID: \"d59afa61-6c33-432d-822a-da4a1c6519bf\") " pod="openstack/nova-metadata-0" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.276391 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d59afa61-6c33-432d-822a-da4a1c6519bf-logs\") pod \"nova-metadata-0\" (UID: \"d59afa61-6c33-432d-822a-da4a1c6519bf\") " pod="openstack/nova-metadata-0" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.276453 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d59afa61-6c33-432d-822a-da4a1c6519bf-config-data\") pod \"nova-metadata-0\" (UID: \"d59afa61-6c33-432d-822a-da4a1c6519bf\") " pod="openstack/nova-metadata-0" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.276504 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d59afa61-6c33-432d-822a-da4a1c6519bf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d59afa61-6c33-432d-822a-da4a1c6519bf\") " pod="openstack/nova-metadata-0" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.276644 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d59afa61-6c33-432d-822a-da4a1c6519bf-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d59afa61-6c33-432d-822a-da4a1c6519bf\") " pod="openstack/nova-metadata-0" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.379069 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d59afa61-6c33-432d-822a-da4a1c6519bf-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d59afa61-6c33-432d-822a-da4a1c6519bf\") " pod="openstack/nova-metadata-0" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.379150 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk8f4\" (UniqueName: \"kubernetes.io/projected/d59afa61-6c33-432d-822a-da4a1c6519bf-kube-api-access-pk8f4\") pod \"nova-metadata-0\" (UID: \"d59afa61-6c33-432d-822a-da4a1c6519bf\") " pod="openstack/nova-metadata-0" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.379194 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d59afa61-6c33-432d-822a-da4a1c6519bf-logs\") pod \"nova-metadata-0\" (UID: \"d59afa61-6c33-432d-822a-da4a1c6519bf\") " pod="openstack/nova-metadata-0" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.379240 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d59afa61-6c33-432d-822a-da4a1c6519bf-config-data\") pod \"nova-metadata-0\" (UID: \"d59afa61-6c33-432d-822a-da4a1c6519bf\") " pod="openstack/nova-metadata-0" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.379284 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d59afa61-6c33-432d-822a-da4a1c6519bf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d59afa61-6c33-432d-822a-da4a1c6519bf\") " pod="openstack/nova-metadata-0" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.379880 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d59afa61-6c33-432d-822a-da4a1c6519bf-logs\") pod \"nova-metadata-0\" (UID: \"d59afa61-6c33-432d-822a-da4a1c6519bf\") " pod="openstack/nova-metadata-0" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.385708 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d59afa61-6c33-432d-822a-da4a1c6519bf-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"d59afa61-6c33-432d-822a-da4a1c6519bf\") " pod="openstack/nova-metadata-0" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.385916 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d59afa61-6c33-432d-822a-da4a1c6519bf-config-data\") pod \"nova-metadata-0\" (UID: \"d59afa61-6c33-432d-822a-da4a1c6519bf\") " pod="openstack/nova-metadata-0" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.386150 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d59afa61-6c33-432d-822a-da4a1c6519bf-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"d59afa61-6c33-432d-822a-da4a1c6519bf\") " pod="openstack/nova-metadata-0" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.396416 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk8f4\" (UniqueName: \"kubernetes.io/projected/d59afa61-6c33-432d-822a-da4a1c6519bf-kube-api-access-pk8f4\") pod \"nova-metadata-0\" (UID: \"d59afa61-6c33-432d-822a-da4a1c6519bf\") " pod="openstack/nova-metadata-0" Mar 17 09:33:09 crc kubenswrapper[4813]: I0317 09:33:09.520364 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 09:33:10 crc kubenswrapper[4813]: I0317 09:33:10.040909 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 09:33:10 crc kubenswrapper[4813]: I0317 09:33:10.746205 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e" path="/var/lib/kubelet/pods/5b6fe5fb-6ea8-4b5c-96d1-68148dbec03e/volumes" Mar 17 09:33:10 crc kubenswrapper[4813]: I0317 09:33:10.831505 4813 generic.go:334] "Generic (PLEG): container finished" podID="500133a3-3767-4bd8-8011-066d1b77ddca" containerID="3671bbe1c1f3b2c6bde24a80e90ff0a32ab448a617d0b0d5858aeb5ce6fe5c63" exitCode=0 Mar 17 09:33:10 crc kubenswrapper[4813]: I0317 09:33:10.831656 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vqcqs" event={"ID":"500133a3-3767-4bd8-8011-066d1b77ddca","Type":"ContainerDied","Data":"3671bbe1c1f3b2c6bde24a80e90ff0a32ab448a617d0b0d5858aeb5ce6fe5c63"} Mar 17 09:33:10 crc kubenswrapper[4813]: I0317 09:33:10.833666 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d59afa61-6c33-432d-822a-da4a1c6519bf","Type":"ContainerStarted","Data":"bf1c1b27bee41b6ef2e189bee878b3138de7ea8a5ba83d7c5bdbca2c59c972e5"} Mar 17 09:33:10 crc kubenswrapper[4813]: I0317 09:33:10.833716 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d59afa61-6c33-432d-822a-da4a1c6519bf","Type":"ContainerStarted","Data":"c84ddc4f4b54e2d9b7e52d0b46a8d002c4d98741d9b10c39325cdc0f03d731e3"} Mar 17 09:33:10 crc kubenswrapper[4813]: I0317 09:33:10.833726 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d59afa61-6c33-432d-822a-da4a1c6519bf","Type":"ContainerStarted","Data":"214b167a4cfd5dc6e08be5ce4a5e09f66053ece7047d5a250273c694252f641a"} Mar 17 09:33:10 crc kubenswrapper[4813]: I0317 09:33:10.867015 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=1.8669929760000001 podStartE2EDuration="1.866992976s" podCreationTimestamp="2026-03-17 09:33:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:33:10.857068675 +0000 UTC m=+1412.957872174" watchObservedRunningTime="2026-03-17 09:33:10.866992976 +0000 UTC m=+1412.967796475" Mar 17 09:33:11 crc kubenswrapper[4813]: I0317 09:33:11.106103 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 17 09:33:11 crc kubenswrapper[4813]: I0317 09:33:11.845727 4813 generic.go:334] "Generic (PLEG): container finished" podID="d08fdcd8-7f6c-468e-bc05-5f0c8a4add79" containerID="ebdf2c0cb40a2d460d25104c47216d741f61c3ee45bfae8c282b3ef6465331d1" exitCode=0 Mar 17 09:33:11 crc kubenswrapper[4813]: I0317 09:33:11.845861 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2j8vj" event={"ID":"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79","Type":"ContainerDied","Data":"ebdf2c0cb40a2d460d25104c47216d741f61c3ee45bfae8c282b3ef6465331d1"} Mar 17 09:33:12 crc kubenswrapper[4813]: I0317 09:33:12.289927 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vqcqs" Mar 17 09:33:12 crc kubenswrapper[4813]: I0317 09:33:12.343665 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/500133a3-3767-4bd8-8011-066d1b77ddca-combined-ca-bundle\") pod \"500133a3-3767-4bd8-8011-066d1b77ddca\" (UID: \"500133a3-3767-4bd8-8011-066d1b77ddca\") " Mar 17 09:33:12 crc kubenswrapper[4813]: I0317 09:33:12.345857 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/500133a3-3767-4bd8-8011-066d1b77ddca-scripts\") pod \"500133a3-3767-4bd8-8011-066d1b77ddca\" (UID: \"500133a3-3767-4bd8-8011-066d1b77ddca\") " Mar 17 09:33:12 crc kubenswrapper[4813]: I0317 09:33:12.346157 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/500133a3-3767-4bd8-8011-066d1b77ddca-config-data\") pod \"500133a3-3767-4bd8-8011-066d1b77ddca\" (UID: \"500133a3-3767-4bd8-8011-066d1b77ddca\") " Mar 17 09:33:12 crc kubenswrapper[4813]: I0317 09:33:12.346375 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6v7d\" (UniqueName: \"kubernetes.io/projected/500133a3-3767-4bd8-8011-066d1b77ddca-kube-api-access-b6v7d\") pod \"500133a3-3767-4bd8-8011-066d1b77ddca\" (UID: \"500133a3-3767-4bd8-8011-066d1b77ddca\") " Mar 17 09:33:12 crc kubenswrapper[4813]: I0317 09:33:12.353456 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/500133a3-3767-4bd8-8011-066d1b77ddca-kube-api-access-b6v7d" (OuterVolumeSpecName: "kube-api-access-b6v7d") pod "500133a3-3767-4bd8-8011-066d1b77ddca" (UID: "500133a3-3767-4bd8-8011-066d1b77ddca"). InnerVolumeSpecName "kube-api-access-b6v7d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:33:12 crc kubenswrapper[4813]: I0317 09:33:12.369289 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/500133a3-3767-4bd8-8011-066d1b77ddca-scripts" (OuterVolumeSpecName: "scripts") pod "500133a3-3767-4bd8-8011-066d1b77ddca" (UID: "500133a3-3767-4bd8-8011-066d1b77ddca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:12 crc kubenswrapper[4813]: I0317 09:33:12.404001 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/500133a3-3767-4bd8-8011-066d1b77ddca-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "500133a3-3767-4bd8-8011-066d1b77ddca" (UID: "500133a3-3767-4bd8-8011-066d1b77ddca"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:12 crc kubenswrapper[4813]: I0317 09:33:12.406217 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/500133a3-3767-4bd8-8011-066d1b77ddca-config-data" (OuterVolumeSpecName: "config-data") pod "500133a3-3767-4bd8-8011-066d1b77ddca" (UID: "500133a3-3767-4bd8-8011-066d1b77ddca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:12 crc kubenswrapper[4813]: I0317 09:33:12.449070 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/500133a3-3767-4bd8-8011-066d1b77ddca-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:12 crc kubenswrapper[4813]: I0317 09:33:12.449100 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/500133a3-3767-4bd8-8011-066d1b77ddca-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:12 crc kubenswrapper[4813]: I0317 09:33:12.449109 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/500133a3-3767-4bd8-8011-066d1b77ddca-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:12 crc kubenswrapper[4813]: I0317 09:33:12.449116 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6v7d\" (UniqueName: \"kubernetes.io/projected/500133a3-3767-4bd8-8011-066d1b77ddca-kube-api-access-b6v7d\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:12 crc kubenswrapper[4813]: I0317 09:33:12.861953 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-vqcqs" Mar 17 09:33:12 crc kubenswrapper[4813]: I0317 09:33:12.863259 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-vqcqs" event={"ID":"500133a3-3767-4bd8-8011-066d1b77ddca","Type":"ContainerDied","Data":"7324d982259930e6c1d95232e3056e4ef1b9d2d2d6a9f01a3b30b57066151a32"} Mar 17 09:33:12 crc kubenswrapper[4813]: I0317 09:33:12.863356 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7324d982259930e6c1d95232e3056e4ef1b9d2d2d6a9f01a3b30b57066151a32" Mar 17 09:33:12 crc kubenswrapper[4813]: I0317 09:33:12.978091 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 17 09:33:12 crc kubenswrapper[4813]: E0317 09:33:12.978721 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="500133a3-3767-4bd8-8011-066d1b77ddca" containerName="nova-cell1-conductor-db-sync" Mar 17 09:33:12 crc kubenswrapper[4813]: I0317 09:33:12.978750 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="500133a3-3767-4bd8-8011-066d1b77ddca" containerName="nova-cell1-conductor-db-sync" Mar 17 09:33:12 crc kubenswrapper[4813]: I0317 09:33:12.979177 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="500133a3-3767-4bd8-8011-066d1b77ddca" containerName="nova-cell1-conductor-db-sync" Mar 17 09:33:12 crc kubenswrapper[4813]: I0317 09:33:12.980218 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 17 09:33:12 crc kubenswrapper[4813]: I0317 09:33:12.983065 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.004667 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.056595 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.062762 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8cfa16e-a939-4e80-9e09-2a632e7bd29e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c8cfa16e-a939-4e80-9e09-2a632e7bd29e\") " pod="openstack/nova-cell1-conductor-0" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.062911 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnk9s\" (UniqueName: \"kubernetes.io/projected/c8cfa16e-a939-4e80-9e09-2a632e7bd29e-kube-api-access-xnk9s\") pod \"nova-cell1-conductor-0\" (UID: \"c8cfa16e-a939-4e80-9e09-2a632e7bd29e\") " pod="openstack/nova-cell1-conductor-0" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.063242 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8cfa16e-a939-4e80-9e09-2a632e7bd29e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c8cfa16e-a939-4e80-9e09-2a632e7bd29e\") " pod="openstack/nova-cell1-conductor-0" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.096231 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.139034 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.139359 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.164867 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8cfa16e-a939-4e80-9e09-2a632e7bd29e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c8cfa16e-a939-4e80-9e09-2a632e7bd29e\") " pod="openstack/nova-cell1-conductor-0" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.164992 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8cfa16e-a939-4e80-9e09-2a632e7bd29e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c8cfa16e-a939-4e80-9e09-2a632e7bd29e\") " pod="openstack/nova-cell1-conductor-0" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.165090 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnk9s\" (UniqueName: \"kubernetes.io/projected/c8cfa16e-a939-4e80-9e09-2a632e7bd29e-kube-api-access-xnk9s\") pod \"nova-cell1-conductor-0\" (UID: \"c8cfa16e-a939-4e80-9e09-2a632e7bd29e\") " pod="openstack/nova-cell1-conductor-0" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.171330 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c8cfa16e-a939-4e80-9e09-2a632e7bd29e-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"c8cfa16e-a939-4e80-9e09-2a632e7bd29e\") " pod="openstack/nova-cell1-conductor-0" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.171376 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c8cfa16e-a939-4e80-9e09-2a632e7bd29e-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"c8cfa16e-a939-4e80-9e09-2a632e7bd29e\") " pod="openstack/nova-cell1-conductor-0" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.182779 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnk9s\" (UniqueName: \"kubernetes.io/projected/c8cfa16e-a939-4e80-9e09-2a632e7bd29e-kube-api-access-xnk9s\") pod \"nova-cell1-conductor-0\" (UID: \"c8cfa16e-a939-4e80-9e09-2a632e7bd29e\") " pod="openstack/nova-cell1-conductor-0" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.319158 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.340280 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2j8vj" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.368035 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-combined-ca-bundle\") pod \"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79\" (UID: \"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79\") " Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.368199 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-scripts\") pod \"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79\" (UID: \"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79\") " Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.368309 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n5dw\" (UniqueName: \"kubernetes.io/projected/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-kube-api-access-6n5dw\") pod \"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79\" (UID: \"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79\") " Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.368373 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-config-data\") pod \"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79\" (UID: \"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79\") " Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.375148 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.416544 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-scripts" (OuterVolumeSpecName: "scripts") pod "d08fdcd8-7f6c-468e-bc05-5f0c8a4add79" (UID: "d08fdcd8-7f6c-468e-bc05-5f0c8a4add79"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.418310 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-kube-api-access-6n5dw" (OuterVolumeSpecName: "kube-api-access-6n5dw") pod "d08fdcd8-7f6c-468e-bc05-5f0c8a4add79" (UID: "d08fdcd8-7f6c-468e-bc05-5f0c8a4add79"). InnerVolumeSpecName "kube-api-access-6n5dw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.448913 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-config-data" (OuterVolumeSpecName: "config-data") pod "d08fdcd8-7f6c-468e-bc05-5f0c8a4add79" (UID: "d08fdcd8-7f6c-468e-bc05-5f0c8a4add79"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.464710 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d08fdcd8-7f6c-468e-bc05-5f0c8a4add79" (UID: "d08fdcd8-7f6c-468e-bc05-5f0c8a4add79"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.471200 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6n5dw\" (UniqueName: \"kubernetes.io/projected/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-kube-api-access-6n5dw\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.471226 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.471237 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.471244 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.482748 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-7wdg7"] Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.483042 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" podUID="899c033d-3f0f-450b-8c73-227886dc17e0" containerName="dnsmasq-dns" containerID="cri-o://4d02f3206c8ca2e0e95ae4f6269f84d8d04299fb1615e599a5f7a19d1ec7a1ad" gracePeriod=10 Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.864227 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.879731 4813 generic.go:334] "Generic (PLEG): container finished" podID="899c033d-3f0f-450b-8c73-227886dc17e0" containerID="4d02f3206c8ca2e0e95ae4f6269f84d8d04299fb1615e599a5f7a19d1ec7a1ad" exitCode=0 Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.880022 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" event={"ID":"899c033d-3f0f-450b-8c73-227886dc17e0","Type":"ContainerDied","Data":"4d02f3206c8ca2e0e95ae4f6269f84d8d04299fb1615e599a5f7a19d1ec7a1ad"} Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.884866 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-2j8vj" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.889811 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-2j8vj" event={"ID":"d08fdcd8-7f6c-468e-bc05-5f0c8a4add79","Type":"ContainerDied","Data":"6094dff23473783db9834ac407f17f80a71a7c2141c3fe7d0ae6ed4f4c911da8"} Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.889836 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6094dff23473783db9834ac407f17f80a71a7c2141c3fe7d0ae6ed4f4c911da8" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.929829 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.967809 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.985430 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-ovsdbserver-sb\") pod \"899c033d-3f0f-450b-8c73-227886dc17e0\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.985616 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnrjq\" (UniqueName: \"kubernetes.io/projected/899c033d-3f0f-450b-8c73-227886dc17e0-kube-api-access-fnrjq\") pod \"899c033d-3f0f-450b-8c73-227886dc17e0\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.985786 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-dns-swift-storage-0\") pod \"899c033d-3f0f-450b-8c73-227886dc17e0\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.985946 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-dns-svc\") pod \"899c033d-3f0f-450b-8c73-227886dc17e0\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.986301 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-ovsdbserver-nb\") pod \"899c033d-3f0f-450b-8c73-227886dc17e0\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " Mar 17 09:33:13 crc kubenswrapper[4813]: I0317 09:33:13.986399 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-config\") pod \"899c033d-3f0f-450b-8c73-227886dc17e0\" (UID: \"899c033d-3f0f-450b-8c73-227886dc17e0\") " Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:13.999647 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/899c033d-3f0f-450b-8c73-227886dc17e0-kube-api-access-fnrjq" (OuterVolumeSpecName: "kube-api-access-fnrjq") pod "899c033d-3f0f-450b-8c73-227886dc17e0" (UID: "899c033d-3f0f-450b-8c73-227886dc17e0"). InnerVolumeSpecName "kube-api-access-fnrjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.073689 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.073954 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d318c13f-5f03-40af-86a1-4fd58b6f6b70" containerName="nova-api-log" containerID="cri-o://96b00c2c3649f24a02186e3513367217ea4a0c7995fc302aaf961370b6f222d6" gracePeriod=30 Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.074181 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="d318c13f-5f03-40af-86a1-4fd58b6f6b70" containerName="nova-api-api" containerID="cri-o://997adf8213b3dd1c07c11b89cc0069428f4f4ee873ac17011cab5fe52d0b58a7" gracePeriod=30 Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.084977 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.085246 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d59afa61-6c33-432d-822a-da4a1c6519bf" containerName="nova-metadata-log" containerID="cri-o://c84ddc4f4b54e2d9b7e52d0b46a8d002c4d98741d9b10c39325cdc0f03d731e3" gracePeriod=30 Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.085430 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="d59afa61-6c33-432d-822a-da4a1c6519bf" containerName="nova-metadata-metadata" containerID="cri-o://bf1c1b27bee41b6ef2e189bee878b3138de7ea8a5ba83d7c5bdbca2c59c972e5" gracePeriod=30 Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.089383 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d318c13f-5f03-40af-86a1-4fd58b6f6b70" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": EOF" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.089675 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnrjq\" (UniqueName: \"kubernetes.io/projected/899c033d-3f0f-450b-8c73-227886dc17e0-kube-api-access-fnrjq\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.089580 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="d318c13f-5f03-40af-86a1-4fd58b6f6b70" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.195:8774/\": EOF" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.113898 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.113947 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.121502 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "899c033d-3f0f-450b-8c73-227886dc17e0" (UID: "899c033d-3f0f-450b-8c73-227886dc17e0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.144856 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "899c033d-3f0f-450b-8c73-227886dc17e0" (UID: "899c033d-3f0f-450b-8c73-227886dc17e0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.148765 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-config" (OuterVolumeSpecName: "config") pod "899c033d-3f0f-450b-8c73-227886dc17e0" (UID: "899c033d-3f0f-450b-8c73-227886dc17e0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.151881 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "899c033d-3f0f-450b-8c73-227886dc17e0" (UID: "899c033d-3f0f-450b-8c73-227886dc17e0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.160501 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "899c033d-3f0f-450b-8c73-227886dc17e0" (UID: "899c033d-3f0f-450b-8c73-227886dc17e0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.191173 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.191460 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.191522 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.191579 4813 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.191653 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/899c033d-3f0f-450b-8c73-227886dc17e0-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.551520 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.740351 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.806392 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pk8f4\" (UniqueName: \"kubernetes.io/projected/d59afa61-6c33-432d-822a-da4a1c6519bf-kube-api-access-pk8f4\") pod \"d59afa61-6c33-432d-822a-da4a1c6519bf\" (UID: \"d59afa61-6c33-432d-822a-da4a1c6519bf\") " Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.806737 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d59afa61-6c33-432d-822a-da4a1c6519bf-config-data\") pod \"d59afa61-6c33-432d-822a-da4a1c6519bf\" (UID: \"d59afa61-6c33-432d-822a-da4a1c6519bf\") " Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.806808 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d59afa61-6c33-432d-822a-da4a1c6519bf-logs\") pod \"d59afa61-6c33-432d-822a-da4a1c6519bf\" (UID: \"d59afa61-6c33-432d-822a-da4a1c6519bf\") " Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.806991 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d59afa61-6c33-432d-822a-da4a1c6519bf-nova-metadata-tls-certs\") pod \"d59afa61-6c33-432d-822a-da4a1c6519bf\" (UID: \"d59afa61-6c33-432d-822a-da4a1c6519bf\") " Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.807064 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d59afa61-6c33-432d-822a-da4a1c6519bf-combined-ca-bundle\") pod \"d59afa61-6c33-432d-822a-da4a1c6519bf\" (UID: \"d59afa61-6c33-432d-822a-da4a1c6519bf\") " Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.808291 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d59afa61-6c33-432d-822a-da4a1c6519bf-logs" (OuterVolumeSpecName: "logs") pod "d59afa61-6c33-432d-822a-da4a1c6519bf" (UID: "d59afa61-6c33-432d-822a-da4a1c6519bf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.835286 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d59afa61-6c33-432d-822a-da4a1c6519bf-kube-api-access-pk8f4" (OuterVolumeSpecName: "kube-api-access-pk8f4") pod "d59afa61-6c33-432d-822a-da4a1c6519bf" (UID: "d59afa61-6c33-432d-822a-da4a1c6519bf"). InnerVolumeSpecName "kube-api-access-pk8f4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.851409 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d59afa61-6c33-432d-822a-da4a1c6519bf-config-data" (OuterVolumeSpecName: "config-data") pod "d59afa61-6c33-432d-822a-da4a1c6519bf" (UID: "d59afa61-6c33-432d-822a-da4a1c6519bf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.864363 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d59afa61-6c33-432d-822a-da4a1c6519bf-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "d59afa61-6c33-432d-822a-da4a1c6519bf" (UID: "d59afa61-6c33-432d-822a-da4a1c6519bf"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.869813 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d59afa61-6c33-432d-822a-da4a1c6519bf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d59afa61-6c33-432d-822a-da4a1c6519bf" (UID: "d59afa61-6c33-432d-822a-da4a1c6519bf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.896751 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" event={"ID":"899c033d-3f0f-450b-8c73-227886dc17e0","Type":"ContainerDied","Data":"aeb0257f37af242b6faeda2ee02ffd22232fe3a8216b2dcaa2b1a8a9387c0faf"} Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.896892 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5784cf869f-7wdg7" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.897162 4813 scope.go:117] "RemoveContainer" containerID="4d02f3206c8ca2e0e95ae4f6269f84d8d04299fb1615e599a5f7a19d1ec7a1ad" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.899929 4813 generic.go:334] "Generic (PLEG): container finished" podID="d59afa61-6c33-432d-822a-da4a1c6519bf" containerID="bf1c1b27bee41b6ef2e189bee878b3138de7ea8a5ba83d7c5bdbca2c59c972e5" exitCode=0 Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.899951 4813 generic.go:334] "Generic (PLEG): container finished" podID="d59afa61-6c33-432d-822a-da4a1c6519bf" containerID="c84ddc4f4b54e2d9b7e52d0b46a8d002c4d98741d9b10c39325cdc0f03d731e3" exitCode=143 Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.899994 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d59afa61-6c33-432d-822a-da4a1c6519bf","Type":"ContainerDied","Data":"bf1c1b27bee41b6ef2e189bee878b3138de7ea8a5ba83d7c5bdbca2c59c972e5"} Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.900019 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d59afa61-6c33-432d-822a-da4a1c6519bf","Type":"ContainerDied","Data":"c84ddc4f4b54e2d9b7e52d0b46a8d002c4d98741d9b10c39325cdc0f03d731e3"} Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.900029 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"d59afa61-6c33-432d-822a-da4a1c6519bf","Type":"ContainerDied","Data":"214b167a4cfd5dc6e08be5ce4a5e09f66053ece7047d5a250273c694252f641a"} Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.900558 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.902575 4813 generic.go:334] "Generic (PLEG): container finished" podID="d318c13f-5f03-40af-86a1-4fd58b6f6b70" containerID="96b00c2c3649f24a02186e3513367217ea4a0c7995fc302aaf961370b6f222d6" exitCode=143 Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.902644 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d318c13f-5f03-40af-86a1-4fd58b6f6b70","Type":"ContainerDied","Data":"96b00c2c3649f24a02186e3513367217ea4a0c7995fc302aaf961370b6f222d6"} Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.905045 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c8cfa16e-a939-4e80-9e09-2a632e7bd29e","Type":"ContainerStarted","Data":"14a9af492f5bb45b8989e48ef1ebfea9c744ea3b7b4d2791c4518d3f3f3a6ab9"} Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.905143 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"c8cfa16e-a939-4e80-9e09-2a632e7bd29e","Type":"ContainerStarted","Data":"9895d9d4d8d5a9e55326b7d3cb8642d1ff7346d81768d9861b66f5f3c52eeb05"} Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.905330 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.908923 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d59afa61-6c33-432d-822a-da4a1c6519bf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.908949 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pk8f4\" (UniqueName: \"kubernetes.io/projected/d59afa61-6c33-432d-822a-da4a1c6519bf-kube-api-access-pk8f4\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.908962 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d59afa61-6c33-432d-822a-da4a1c6519bf-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.908971 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d59afa61-6c33-432d-822a-da4a1c6519bf-logs\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.908981 4813 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/d59afa61-6c33-432d-822a-da4a1c6519bf-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.928376 4813 scope.go:117] "RemoveContainer" containerID="07fc41d3607cf1234be698707cf2db9ace33c8187811d985db0d032100082ac8" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.929183 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-7wdg7"] Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.951011 4813 scope.go:117] "RemoveContainer" containerID="bf1c1b27bee41b6ef2e189bee878b3138de7ea8a5ba83d7c5bdbca2c59c972e5" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.953385 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5784cf869f-7wdg7"] Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.954411 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=2.954394418 podStartE2EDuration="2.954394418s" podCreationTimestamp="2026-03-17 09:33:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:33:14.932207521 +0000 UTC m=+1417.033011020" watchObservedRunningTime="2026-03-17 09:33:14.954394418 +0000 UTC m=+1417.055197917" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.979300 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.982851 4813 scope.go:117] "RemoveContainer" containerID="c84ddc4f4b54e2d9b7e52d0b46a8d002c4d98741d9b10c39325cdc0f03d731e3" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.987352 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.996941 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 17 09:33:14 crc kubenswrapper[4813]: E0317 09:33:14.997655 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d59afa61-6c33-432d-822a-da4a1c6519bf" containerName="nova-metadata-log" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.997681 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d59afa61-6c33-432d-822a-da4a1c6519bf" containerName="nova-metadata-log" Mar 17 09:33:14 crc kubenswrapper[4813]: E0317 09:33:14.997707 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d59afa61-6c33-432d-822a-da4a1c6519bf" containerName="nova-metadata-metadata" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.997715 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d59afa61-6c33-432d-822a-da4a1c6519bf" containerName="nova-metadata-metadata" Mar 17 09:33:14 crc kubenswrapper[4813]: E0317 09:33:14.997738 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="899c033d-3f0f-450b-8c73-227886dc17e0" containerName="init" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.997745 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="899c033d-3f0f-450b-8c73-227886dc17e0" containerName="init" Mar 17 09:33:14 crc kubenswrapper[4813]: E0317 09:33:14.997756 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d08fdcd8-7f6c-468e-bc05-5f0c8a4add79" containerName="nova-manage" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.997766 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d08fdcd8-7f6c-468e-bc05-5f0c8a4add79" containerName="nova-manage" Mar 17 09:33:14 crc kubenswrapper[4813]: E0317 09:33:14.997841 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="899c033d-3f0f-450b-8c73-227886dc17e0" containerName="dnsmasq-dns" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.997852 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="899c033d-3f0f-450b-8c73-227886dc17e0" containerName="dnsmasq-dns" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.998153 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d08fdcd8-7f6c-468e-bc05-5f0c8a4add79" containerName="nova-manage" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.998165 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d59afa61-6c33-432d-822a-da4a1c6519bf" containerName="nova-metadata-metadata" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.998181 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d59afa61-6c33-432d-822a-da4a1c6519bf" containerName="nova-metadata-log" Mar 17 09:33:14 crc kubenswrapper[4813]: I0317 09:33:14.998188 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="899c033d-3f0f-450b-8c73-227886dc17e0" containerName="dnsmasq-dns" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.001474 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.008811 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.013564 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.014196 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.016024 4813 scope.go:117] "RemoveContainer" containerID="bf1c1b27bee41b6ef2e189bee878b3138de7ea8a5ba83d7c5bdbca2c59c972e5" Mar 17 09:33:15 crc kubenswrapper[4813]: E0317 09:33:15.016348 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf1c1b27bee41b6ef2e189bee878b3138de7ea8a5ba83d7c5bdbca2c59c972e5\": container with ID starting with bf1c1b27bee41b6ef2e189bee878b3138de7ea8a5ba83d7c5bdbca2c59c972e5 not found: ID does not exist" containerID="bf1c1b27bee41b6ef2e189bee878b3138de7ea8a5ba83d7c5bdbca2c59c972e5" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.016372 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf1c1b27bee41b6ef2e189bee878b3138de7ea8a5ba83d7c5bdbca2c59c972e5"} err="failed to get container status \"bf1c1b27bee41b6ef2e189bee878b3138de7ea8a5ba83d7c5bdbca2c59c972e5\": rpc error: code = NotFound desc = could not find container \"bf1c1b27bee41b6ef2e189bee878b3138de7ea8a5ba83d7c5bdbca2c59c972e5\": container with ID starting with bf1c1b27bee41b6ef2e189bee878b3138de7ea8a5ba83d7c5bdbca2c59c972e5 not found: ID does not exist" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.016391 4813 scope.go:117] "RemoveContainer" containerID="c84ddc4f4b54e2d9b7e52d0b46a8d002c4d98741d9b10c39325cdc0f03d731e3" Mar 17 09:33:15 crc kubenswrapper[4813]: E0317 09:33:15.016561 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c84ddc4f4b54e2d9b7e52d0b46a8d002c4d98741d9b10c39325cdc0f03d731e3\": container with ID starting with c84ddc4f4b54e2d9b7e52d0b46a8d002c4d98741d9b10c39325cdc0f03d731e3 not found: ID does not exist" containerID="c84ddc4f4b54e2d9b7e52d0b46a8d002c4d98741d9b10c39325cdc0f03d731e3" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.016575 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c84ddc4f4b54e2d9b7e52d0b46a8d002c4d98741d9b10c39325cdc0f03d731e3"} err="failed to get container status \"c84ddc4f4b54e2d9b7e52d0b46a8d002c4d98741d9b10c39325cdc0f03d731e3\": rpc error: code = NotFound desc = could not find container \"c84ddc4f4b54e2d9b7e52d0b46a8d002c4d98741d9b10c39325cdc0f03d731e3\": container with ID starting with c84ddc4f4b54e2d9b7e52d0b46a8d002c4d98741d9b10c39325cdc0f03d731e3 not found: ID does not exist" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.016586 4813 scope.go:117] "RemoveContainer" containerID="bf1c1b27bee41b6ef2e189bee878b3138de7ea8a5ba83d7c5bdbca2c59c972e5" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.017244 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf1c1b27bee41b6ef2e189bee878b3138de7ea8a5ba83d7c5bdbca2c59c972e5"} err="failed to get container status \"bf1c1b27bee41b6ef2e189bee878b3138de7ea8a5ba83d7c5bdbca2c59c972e5\": rpc error: code = NotFound desc = could not find container \"bf1c1b27bee41b6ef2e189bee878b3138de7ea8a5ba83d7c5bdbca2c59c972e5\": container with ID starting with bf1c1b27bee41b6ef2e189bee878b3138de7ea8a5ba83d7c5bdbca2c59c972e5 not found: ID does not exist" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.017259 4813 scope.go:117] "RemoveContainer" containerID="c84ddc4f4b54e2d9b7e52d0b46a8d002c4d98741d9b10c39325cdc0f03d731e3" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.017410 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c84ddc4f4b54e2d9b7e52d0b46a8d002c4d98741d9b10c39325cdc0f03d731e3"} err="failed to get container status \"c84ddc4f4b54e2d9b7e52d0b46a8d002c4d98741d9b10c39325cdc0f03d731e3\": rpc error: code = NotFound desc = could not find container \"c84ddc4f4b54e2d9b7e52d0b46a8d002c4d98741d9b10c39325cdc0f03d731e3\": container with ID starting with c84ddc4f4b54e2d9b7e52d0b46a8d002c4d98741d9b10c39325cdc0f03d731e3 not found: ID does not exist" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.113653 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\") " pod="openstack/nova-metadata-0" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.113962 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-logs\") pod \"nova-metadata-0\" (UID: \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\") " pod="openstack/nova-metadata-0" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.114294 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87qrq\" (UniqueName: \"kubernetes.io/projected/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-kube-api-access-87qrq\") pod \"nova-metadata-0\" (UID: \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\") " pod="openstack/nova-metadata-0" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.114455 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\") " pod="openstack/nova-metadata-0" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.114574 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-config-data\") pod \"nova-metadata-0\" (UID: \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\") " pod="openstack/nova-metadata-0" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.216367 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\") " pod="openstack/nova-metadata-0" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.216429 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-logs\") pod \"nova-metadata-0\" (UID: \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\") " pod="openstack/nova-metadata-0" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.216471 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87qrq\" (UniqueName: \"kubernetes.io/projected/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-kube-api-access-87qrq\") pod \"nova-metadata-0\" (UID: \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\") " pod="openstack/nova-metadata-0" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.216491 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\") " pod="openstack/nova-metadata-0" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.216523 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-config-data\") pod \"nova-metadata-0\" (UID: \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\") " pod="openstack/nova-metadata-0" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.217082 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-logs\") pod \"nova-metadata-0\" (UID: \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\") " pod="openstack/nova-metadata-0" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.221194 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-config-data\") pod \"nova-metadata-0\" (UID: \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\") " pod="openstack/nova-metadata-0" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.221727 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\") " pod="openstack/nova-metadata-0" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.222233 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\") " pod="openstack/nova-metadata-0" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.236217 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87qrq\" (UniqueName: \"kubernetes.io/projected/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-kube-api-access-87qrq\") pod \"nova-metadata-0\" (UID: \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\") " pod="openstack/nova-metadata-0" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.311573 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.311783 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="59209cd3-5db8-4ddc-806e-1db457164c1f" containerName="kube-state-metrics" containerID="cri-o://64f16397599de8972cf374613af13195a1688ec6b89a4186a51966f796c951a4" gracePeriod=30 Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.327116 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.801654 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.823780 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.920434 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"59209cd3-5db8-4ddc-806e-1db457164c1f","Type":"ContainerDied","Data":"64f16397599de8972cf374613af13195a1688ec6b89a4186a51966f796c951a4"} Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.920926 4813 scope.go:117] "RemoveContainer" containerID="64f16397599de8972cf374613af13195a1688ec6b89a4186a51966f796c951a4" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.920405 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.921159 4813 generic.go:334] "Generic (PLEG): container finished" podID="59209cd3-5db8-4ddc-806e-1db457164c1f" containerID="64f16397599de8972cf374613af13195a1688ec6b89a4186a51966f796c951a4" exitCode=2 Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.921339 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"59209cd3-5db8-4ddc-806e-1db457164c1f","Type":"ContainerDied","Data":"ec5ec8ea069dc659eea9cb2219c465524921b7df81d24a13f4509fb998cd1c92"} Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.931446 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe","Type":"ContainerStarted","Data":"928c1595632eaec6df4ebb7df9f1120bf3779a1a79a9d5774416f7f78fc34048"} Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.937424 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcccg\" (UniqueName: \"kubernetes.io/projected/59209cd3-5db8-4ddc-806e-1db457164c1f-kube-api-access-dcccg\") pod \"59209cd3-5db8-4ddc-806e-1db457164c1f\" (UID: \"59209cd3-5db8-4ddc-806e-1db457164c1f\") " Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.939207 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="baca76f9-d625-4532-9315-8be77971153b" containerName="nova-scheduler-scheduler" containerID="cri-o://b7b096e17f5a01a3456741d7e0f29d468e7898533e0b1e825c1bbf1d682b64ed" gracePeriod=30 Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.947063 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59209cd3-5db8-4ddc-806e-1db457164c1f-kube-api-access-dcccg" (OuterVolumeSpecName: "kube-api-access-dcccg") pod "59209cd3-5db8-4ddc-806e-1db457164c1f" (UID: "59209cd3-5db8-4ddc-806e-1db457164c1f"). InnerVolumeSpecName "kube-api-access-dcccg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.950751 4813 scope.go:117] "RemoveContainer" containerID="64f16397599de8972cf374613af13195a1688ec6b89a4186a51966f796c951a4" Mar 17 09:33:15 crc kubenswrapper[4813]: E0317 09:33:15.951251 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"64f16397599de8972cf374613af13195a1688ec6b89a4186a51966f796c951a4\": container with ID starting with 64f16397599de8972cf374613af13195a1688ec6b89a4186a51966f796c951a4 not found: ID does not exist" containerID="64f16397599de8972cf374613af13195a1688ec6b89a4186a51966f796c951a4" Mar 17 09:33:15 crc kubenswrapper[4813]: I0317 09:33:15.951312 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"64f16397599de8972cf374613af13195a1688ec6b89a4186a51966f796c951a4"} err="failed to get container status \"64f16397599de8972cf374613af13195a1688ec6b89a4186a51966f796c951a4\": rpc error: code = NotFound desc = could not find container \"64f16397599de8972cf374613af13195a1688ec6b89a4186a51966f796c951a4\": container with ID starting with 64f16397599de8972cf374613af13195a1688ec6b89a4186a51966f796c951a4 not found: ID does not exist" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.040766 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcccg\" (UniqueName: \"kubernetes.io/projected/59209cd3-5db8-4ddc-806e-1db457164c1f-kube-api-access-dcccg\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.283055 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.295922 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.318436 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Mar 17 09:33:16 crc kubenswrapper[4813]: E0317 09:33:16.318955 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59209cd3-5db8-4ddc-806e-1db457164c1f" containerName="kube-state-metrics" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.318979 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="59209cd3-5db8-4ddc-806e-1db457164c1f" containerName="kube-state-metrics" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.319215 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="59209cd3-5db8-4ddc-806e-1db457164c1f" containerName="kube-state-metrics" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.320007 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.322281 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.322372 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.330957 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.349578 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e83f1d1-1126-49a5-9a24-1694e7616b61-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9e83f1d1-1126-49a5-9a24-1694e7616b61\") " pod="openstack/kube-state-metrics-0" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.349670 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9e83f1d1-1126-49a5-9a24-1694e7616b61-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9e83f1d1-1126-49a5-9a24-1694e7616b61\") " pod="openstack/kube-state-metrics-0" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.349733 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e83f1d1-1126-49a5-9a24-1694e7616b61-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9e83f1d1-1126-49a5-9a24-1694e7616b61\") " pod="openstack/kube-state-metrics-0" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.349795 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xfhx7\" (UniqueName: \"kubernetes.io/projected/9e83f1d1-1126-49a5-9a24-1694e7616b61-kube-api-access-xfhx7\") pod \"kube-state-metrics-0\" (UID: \"9e83f1d1-1126-49a5-9a24-1694e7616b61\") " pod="openstack/kube-state-metrics-0" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.452035 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e83f1d1-1126-49a5-9a24-1694e7616b61-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9e83f1d1-1126-49a5-9a24-1694e7616b61\") " pod="openstack/kube-state-metrics-0" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.452491 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xfhx7\" (UniqueName: \"kubernetes.io/projected/9e83f1d1-1126-49a5-9a24-1694e7616b61-kube-api-access-xfhx7\") pod \"kube-state-metrics-0\" (UID: \"9e83f1d1-1126-49a5-9a24-1694e7616b61\") " pod="openstack/kube-state-metrics-0" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.452690 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e83f1d1-1126-49a5-9a24-1694e7616b61-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9e83f1d1-1126-49a5-9a24-1694e7616b61\") " pod="openstack/kube-state-metrics-0" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.452740 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9e83f1d1-1126-49a5-9a24-1694e7616b61-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9e83f1d1-1126-49a5-9a24-1694e7616b61\") " pod="openstack/kube-state-metrics-0" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.457118 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e83f1d1-1126-49a5-9a24-1694e7616b61-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"9e83f1d1-1126-49a5-9a24-1694e7616b61\") " pod="openstack/kube-state-metrics-0" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.467539 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e83f1d1-1126-49a5-9a24-1694e7616b61-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"9e83f1d1-1126-49a5-9a24-1694e7616b61\") " pod="openstack/kube-state-metrics-0" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.468443 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/9e83f1d1-1126-49a5-9a24-1694e7616b61-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"9e83f1d1-1126-49a5-9a24-1694e7616b61\") " pod="openstack/kube-state-metrics-0" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.478501 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xfhx7\" (UniqueName: \"kubernetes.io/projected/9e83f1d1-1126-49a5-9a24-1694e7616b61-kube-api-access-xfhx7\") pod \"kube-state-metrics-0\" (UID: \"9e83f1d1-1126-49a5-9a24-1694e7616b61\") " pod="openstack/kube-state-metrics-0" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.638167 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.746791 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59209cd3-5db8-4ddc-806e-1db457164c1f" path="/var/lib/kubelet/pods/59209cd3-5db8-4ddc-806e-1db457164c1f/volumes" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.747441 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="899c033d-3f0f-450b-8c73-227886dc17e0" path="/var/lib/kubelet/pods/899c033d-3f0f-450b-8c73-227886dc17e0/volumes" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.748111 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d59afa61-6c33-432d-822a-da4a1c6519bf" path="/var/lib/kubelet/pods/d59afa61-6c33-432d-822a-da4a1c6519bf/volumes" Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.945499 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe","Type":"ContainerStarted","Data":"a2de365165a9e93f392da0a42a918ce0e6cf111a582c5b174f55c44c7d966791"} Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.945546 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe","Type":"ContainerStarted","Data":"a44bfd63612dfab25234155cad9e67aa869107fcf44e75fc72bab3b9bef5eb28"} Mar 17 09:33:16 crc kubenswrapper[4813]: I0317 09:33:16.998182 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.998158376 podStartE2EDuration="2.998158376s" podCreationTimestamp="2026-03-17 09:33:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:33:16.984643471 +0000 UTC m=+1419.085446990" watchObservedRunningTime="2026-03-17 09:33:16.998158376 +0000 UTC m=+1419.098961875" Mar 17 09:33:17 crc kubenswrapper[4813]: I0317 09:33:17.109568 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Mar 17 09:33:17 crc kubenswrapper[4813]: I0317 09:33:17.486403 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:33:17 crc kubenswrapper[4813]: I0317 09:33:17.486747 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f41fef51-25cf-4306-87fa-4f203d7d18e9" containerName="ceilometer-central-agent" containerID="cri-o://82687c8db83081af109ede7b426e5712fbcae4c4e3a3097ca211e27cbf1fe77e" gracePeriod=30 Mar 17 09:33:17 crc kubenswrapper[4813]: I0317 09:33:17.486852 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f41fef51-25cf-4306-87fa-4f203d7d18e9" containerName="ceilometer-notification-agent" containerID="cri-o://01eefccb4733f564717eef9cd02109e444797b79107d8c82d5c9cdf662bdbe39" gracePeriod=30 Mar 17 09:33:17 crc kubenswrapper[4813]: I0317 09:33:17.486836 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f41fef51-25cf-4306-87fa-4f203d7d18e9" containerName="sg-core" containerID="cri-o://0533d88c48b4cf268e85d8b71e73c298620f3f3e352445936593f80fb3c9954d" gracePeriod=30 Mar 17 09:33:17 crc kubenswrapper[4813]: I0317 09:33:17.486920 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f41fef51-25cf-4306-87fa-4f203d7d18e9" containerName="proxy-httpd" containerID="cri-o://a6b6a160c064ec98158a0a506fb892a51ee338a0fbf0fe48426cb9499e167a73" gracePeriod=30 Mar 17 09:33:17 crc kubenswrapper[4813]: I0317 09:33:17.958061 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9e83f1d1-1126-49a5-9a24-1694e7616b61","Type":"ContainerStarted","Data":"5f5c66b2388cc2041364c1978af466d9b16ae7b6a3f65cfc4f5c679a35bce3be"} Mar 17 09:33:17 crc kubenswrapper[4813]: I0317 09:33:17.961638 4813 generic.go:334] "Generic (PLEG): container finished" podID="f41fef51-25cf-4306-87fa-4f203d7d18e9" containerID="a6b6a160c064ec98158a0a506fb892a51ee338a0fbf0fe48426cb9499e167a73" exitCode=0 Mar 17 09:33:17 crc kubenswrapper[4813]: I0317 09:33:17.961661 4813 generic.go:334] "Generic (PLEG): container finished" podID="f41fef51-25cf-4306-87fa-4f203d7d18e9" containerID="0533d88c48b4cf268e85d8b71e73c298620f3f3e352445936593f80fb3c9954d" exitCode=2 Mar 17 09:33:17 crc kubenswrapper[4813]: I0317 09:33:17.961669 4813 generic.go:334] "Generic (PLEG): container finished" podID="f41fef51-25cf-4306-87fa-4f203d7d18e9" containerID="82687c8db83081af109ede7b426e5712fbcae4c4e3a3097ca211e27cbf1fe77e" exitCode=0 Mar 17 09:33:17 crc kubenswrapper[4813]: I0317 09:33:17.961743 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f41fef51-25cf-4306-87fa-4f203d7d18e9","Type":"ContainerDied","Data":"a6b6a160c064ec98158a0a506fb892a51ee338a0fbf0fe48426cb9499e167a73"} Mar 17 09:33:17 crc kubenswrapper[4813]: I0317 09:33:17.961797 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f41fef51-25cf-4306-87fa-4f203d7d18e9","Type":"ContainerDied","Data":"0533d88c48b4cf268e85d8b71e73c298620f3f3e352445936593f80fb3c9954d"} Mar 17 09:33:17 crc kubenswrapper[4813]: I0317 09:33:17.961811 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f41fef51-25cf-4306-87fa-4f203d7d18e9","Type":"ContainerDied","Data":"82687c8db83081af109ede7b426e5712fbcae4c4e3a3097ca211e27cbf1fe77e"} Mar 17 09:33:18 crc kubenswrapper[4813]: E0317 09:33:18.059232 4813 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b7b096e17f5a01a3456741d7e0f29d468e7898533e0b1e825c1bbf1d682b64ed" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 17 09:33:18 crc kubenswrapper[4813]: E0317 09:33:18.061474 4813 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b7b096e17f5a01a3456741d7e0f29d468e7898533e0b1e825c1bbf1d682b64ed" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 17 09:33:18 crc kubenswrapper[4813]: E0317 09:33:18.063876 4813 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="b7b096e17f5a01a3456741d7e0f29d468e7898533e0b1e825c1bbf1d682b64ed" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 17 09:33:18 crc kubenswrapper[4813]: E0317 09:33:18.063934 4813 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="baca76f9-d625-4532-9315-8be77971153b" containerName="nova-scheduler-scheduler" Mar 17 09:33:18 crc kubenswrapper[4813]: I0317 09:33:18.896825 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 17 09:33:18 crc kubenswrapper[4813]: I0317 09:33:18.975543 4813 generic.go:334] "Generic (PLEG): container finished" podID="baca76f9-d625-4532-9315-8be77971153b" containerID="b7b096e17f5a01a3456741d7e0f29d468e7898533e0b1e825c1bbf1d682b64ed" exitCode=0 Mar 17 09:33:18 crc kubenswrapper[4813]: I0317 09:33:18.975611 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"baca76f9-d625-4532-9315-8be77971153b","Type":"ContainerDied","Data":"b7b096e17f5a01a3456741d7e0f29d468e7898533e0b1e825c1bbf1d682b64ed"} Mar 17 09:33:18 crc kubenswrapper[4813]: I0317 09:33:18.975685 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"baca76f9-d625-4532-9315-8be77971153b","Type":"ContainerDied","Data":"3b878f5897f8fb9a0a746d3e5228b91849e593d27460757caf83ab7c1fe680f9"} Mar 17 09:33:18 crc kubenswrapper[4813]: I0317 09:33:18.975717 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 17 09:33:18 crc kubenswrapper[4813]: I0317 09:33:18.975738 4813 scope.go:117] "RemoveContainer" containerID="b7b096e17f5a01a3456741d7e0f29d468e7898533e0b1e825c1bbf1d682b64ed" Mar 17 09:33:18 crc kubenswrapper[4813]: I0317 09:33:18.994436 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baca76f9-d625-4532-9315-8be77971153b-combined-ca-bundle\") pod \"baca76f9-d625-4532-9315-8be77971153b\" (UID: \"baca76f9-d625-4532-9315-8be77971153b\") " Mar 17 09:33:18 crc kubenswrapper[4813]: I0317 09:33:18.994582 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baca76f9-d625-4532-9315-8be77971153b-config-data\") pod \"baca76f9-d625-4532-9315-8be77971153b\" (UID: \"baca76f9-d625-4532-9315-8be77971153b\") " Mar 17 09:33:18 crc kubenswrapper[4813]: I0317 09:33:18.994652 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljfmb\" (UniqueName: \"kubernetes.io/projected/baca76f9-d625-4532-9315-8be77971153b-kube-api-access-ljfmb\") pod \"baca76f9-d625-4532-9315-8be77971153b\" (UID: \"baca76f9-d625-4532-9315-8be77971153b\") " Mar 17 09:33:18 crc kubenswrapper[4813]: I0317 09:33:18.998904 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/baca76f9-d625-4532-9315-8be77971153b-kube-api-access-ljfmb" (OuterVolumeSpecName: "kube-api-access-ljfmb") pod "baca76f9-d625-4532-9315-8be77971153b" (UID: "baca76f9-d625-4532-9315-8be77971153b"). InnerVolumeSpecName "kube-api-access-ljfmb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.019892 4813 scope.go:117] "RemoveContainer" containerID="b7b096e17f5a01a3456741d7e0f29d468e7898533e0b1e825c1bbf1d682b64ed" Mar 17 09:33:19 crc kubenswrapper[4813]: E0317 09:33:19.020331 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7b096e17f5a01a3456741d7e0f29d468e7898533e0b1e825c1bbf1d682b64ed\": container with ID starting with b7b096e17f5a01a3456741d7e0f29d468e7898533e0b1e825c1bbf1d682b64ed not found: ID does not exist" containerID="b7b096e17f5a01a3456741d7e0f29d468e7898533e0b1e825c1bbf1d682b64ed" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.020364 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7b096e17f5a01a3456741d7e0f29d468e7898533e0b1e825c1bbf1d682b64ed"} err="failed to get container status \"b7b096e17f5a01a3456741d7e0f29d468e7898533e0b1e825c1bbf1d682b64ed\": rpc error: code = NotFound desc = could not find container \"b7b096e17f5a01a3456741d7e0f29d468e7898533e0b1e825c1bbf1d682b64ed\": container with ID starting with b7b096e17f5a01a3456741d7e0f29d468e7898533e0b1e825c1bbf1d682b64ed not found: ID does not exist" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.021349 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baca76f9-d625-4532-9315-8be77971153b-config-data" (OuterVolumeSpecName: "config-data") pod "baca76f9-d625-4532-9315-8be77971153b" (UID: "baca76f9-d625-4532-9315-8be77971153b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.023647 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/baca76f9-d625-4532-9315-8be77971153b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "baca76f9-d625-4532-9315-8be77971153b" (UID: "baca76f9-d625-4532-9315-8be77971153b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.096886 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/baca76f9-d625-4532-9315-8be77971153b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.096918 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/baca76f9-d625-4532-9315-8be77971153b-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.096927 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljfmb\" (UniqueName: \"kubernetes.io/projected/baca76f9-d625-4532-9315-8be77971153b-kube-api-access-ljfmb\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.329777 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.350709 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.361937 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 09:33:19 crc kubenswrapper[4813]: E0317 09:33:19.362290 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="baca76f9-d625-4532-9315-8be77971153b" containerName="nova-scheduler-scheduler" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.362303 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="baca76f9-d625-4532-9315-8be77971153b" containerName="nova-scheduler-scheduler" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.362468 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="baca76f9-d625-4532-9315-8be77971153b" containerName="nova-scheduler-scheduler" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.363027 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.365323 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.406701 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.407846 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69e2e92a-5161-4a09-b59d-464cb4490ff0-config-data\") pod \"nova-scheduler-0\" (UID: \"69e2e92a-5161-4a09-b59d-464cb4490ff0\") " pod="openstack/nova-scheduler-0" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.407885 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69e2e92a-5161-4a09-b59d-464cb4490ff0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"69e2e92a-5161-4a09-b59d-464cb4490ff0\") " pod="openstack/nova-scheduler-0" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.407914 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfphb\" (UniqueName: \"kubernetes.io/projected/69e2e92a-5161-4a09-b59d-464cb4490ff0-kube-api-access-lfphb\") pod \"nova-scheduler-0\" (UID: \"69e2e92a-5161-4a09-b59d-464cb4490ff0\") " pod="openstack/nova-scheduler-0" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.519861 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69e2e92a-5161-4a09-b59d-464cb4490ff0-config-data\") pod \"nova-scheduler-0\" (UID: \"69e2e92a-5161-4a09-b59d-464cb4490ff0\") " pod="openstack/nova-scheduler-0" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.519915 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69e2e92a-5161-4a09-b59d-464cb4490ff0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"69e2e92a-5161-4a09-b59d-464cb4490ff0\") " pod="openstack/nova-scheduler-0" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.519948 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfphb\" (UniqueName: \"kubernetes.io/projected/69e2e92a-5161-4a09-b59d-464cb4490ff0-kube-api-access-lfphb\") pod \"nova-scheduler-0\" (UID: \"69e2e92a-5161-4a09-b59d-464cb4490ff0\") " pod="openstack/nova-scheduler-0" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.536850 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69e2e92a-5161-4a09-b59d-464cb4490ff0-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"69e2e92a-5161-4a09-b59d-464cb4490ff0\") " pod="openstack/nova-scheduler-0" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.536923 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69e2e92a-5161-4a09-b59d-464cb4490ff0-config-data\") pod \"nova-scheduler-0\" (UID: \"69e2e92a-5161-4a09-b59d-464cb4490ff0\") " pod="openstack/nova-scheduler-0" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.547159 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfphb\" (UniqueName: \"kubernetes.io/projected/69e2e92a-5161-4a09-b59d-464cb4490ff0-kube-api-access-lfphb\") pod \"nova-scheduler-0\" (UID: \"69e2e92a-5161-4a09-b59d-464cb4490ff0\") " pod="openstack/nova-scheduler-0" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.594575 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.621455 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-scripts\") pod \"f41fef51-25cf-4306-87fa-4f203d7d18e9\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.621564 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jtjn\" (UniqueName: \"kubernetes.io/projected/f41fef51-25cf-4306-87fa-4f203d7d18e9-kube-api-access-2jtjn\") pod \"f41fef51-25cf-4306-87fa-4f203d7d18e9\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.621657 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f41fef51-25cf-4306-87fa-4f203d7d18e9-log-httpd\") pod \"f41fef51-25cf-4306-87fa-4f203d7d18e9\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.621681 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f41fef51-25cf-4306-87fa-4f203d7d18e9-run-httpd\") pod \"f41fef51-25cf-4306-87fa-4f203d7d18e9\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.621700 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-config-data\") pod \"f41fef51-25cf-4306-87fa-4f203d7d18e9\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.621745 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-sg-core-conf-yaml\") pod \"f41fef51-25cf-4306-87fa-4f203d7d18e9\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.621820 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-combined-ca-bundle\") pod \"f41fef51-25cf-4306-87fa-4f203d7d18e9\" (UID: \"f41fef51-25cf-4306-87fa-4f203d7d18e9\") " Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.623115 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f41fef51-25cf-4306-87fa-4f203d7d18e9-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f41fef51-25cf-4306-87fa-4f203d7d18e9" (UID: "f41fef51-25cf-4306-87fa-4f203d7d18e9"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.623470 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f41fef51-25cf-4306-87fa-4f203d7d18e9-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f41fef51-25cf-4306-87fa-4f203d7d18e9" (UID: "f41fef51-25cf-4306-87fa-4f203d7d18e9"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.661167 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f41fef51-25cf-4306-87fa-4f203d7d18e9-kube-api-access-2jtjn" (OuterVolumeSpecName: "kube-api-access-2jtjn") pod "f41fef51-25cf-4306-87fa-4f203d7d18e9" (UID: "f41fef51-25cf-4306-87fa-4f203d7d18e9"). InnerVolumeSpecName "kube-api-access-2jtjn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.664149 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-scripts" (OuterVolumeSpecName: "scripts") pod "f41fef51-25cf-4306-87fa-4f203d7d18e9" (UID: "f41fef51-25cf-4306-87fa-4f203d7d18e9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.668751 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f41fef51-25cf-4306-87fa-4f203d7d18e9" (UID: "f41fef51-25cf-4306-87fa-4f203d7d18e9"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.709232 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f41fef51-25cf-4306-87fa-4f203d7d18e9" (UID: "f41fef51-25cf-4306-87fa-4f203d7d18e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.724160 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.724189 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jtjn\" (UniqueName: \"kubernetes.io/projected/f41fef51-25cf-4306-87fa-4f203d7d18e9-kube-api-access-2jtjn\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.724201 4813 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f41fef51-25cf-4306-87fa-4f203d7d18e9-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.724211 4813 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f41fef51-25cf-4306-87fa-4f203d7d18e9-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.724219 4813 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.724227 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.730190 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.742666 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-config-data" (OuterVolumeSpecName: "config-data") pod "f41fef51-25cf-4306-87fa-4f203d7d18e9" (UID: "f41fef51-25cf-4306-87fa-4f203d7d18e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.827124 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f41fef51-25cf-4306-87fa-4f203d7d18e9-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.988444 4813 generic.go:334] "Generic (PLEG): container finished" podID="f41fef51-25cf-4306-87fa-4f203d7d18e9" containerID="01eefccb4733f564717eef9cd02109e444797b79107d8c82d5c9cdf662bdbe39" exitCode=0 Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.988506 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f41fef51-25cf-4306-87fa-4f203d7d18e9","Type":"ContainerDied","Data":"01eefccb4733f564717eef9cd02109e444797b79107d8c82d5c9cdf662bdbe39"} Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.988533 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f41fef51-25cf-4306-87fa-4f203d7d18e9","Type":"ContainerDied","Data":"938da892e8016b9e2fd28ea9a2f6d04804e63fdd85adaa92b26e1c84b1116673"} Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.988549 4813 scope.go:117] "RemoveContainer" containerID="a6b6a160c064ec98158a0a506fb892a51ee338a0fbf0fe48426cb9499e167a73" Mar 17 09:33:19 crc kubenswrapper[4813]: I0317 09:33:19.988667 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.002932 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"9e83f1d1-1126-49a5-9a24-1694e7616b61","Type":"ContainerStarted","Data":"171982a71ba5b9ecbb8ba41037d5720c783028a025bfd37569cf0558cff94a94"} Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.003175 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.025231 4813 scope.go:117] "RemoveContainer" containerID="0533d88c48b4cf268e85d8b71e73c298620f3f3e352445936593f80fb3c9954d" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.027583 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.040187 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.047459 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=2.3868726430000002 podStartE2EDuration="4.047435929s" podCreationTimestamp="2026-03-17 09:33:16 +0000 UTC" firstStartedPulling="2026-03-17 09:33:17.111420579 +0000 UTC m=+1419.212224078" lastFinishedPulling="2026-03-17 09:33:18.771983875 +0000 UTC m=+1420.872787364" observedRunningTime="2026-03-17 09:33:20.033326636 +0000 UTC m=+1422.134130175" watchObservedRunningTime="2026-03-17 09:33:20.047435929 +0000 UTC m=+1422.148239428" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.060053 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:33:20 crc kubenswrapper[4813]: E0317 09:33:20.060500 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f41fef51-25cf-4306-87fa-4f203d7d18e9" containerName="ceilometer-notification-agent" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.060521 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f41fef51-25cf-4306-87fa-4f203d7d18e9" containerName="ceilometer-notification-agent" Mar 17 09:33:20 crc kubenswrapper[4813]: E0317 09:33:20.060535 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f41fef51-25cf-4306-87fa-4f203d7d18e9" containerName="ceilometer-central-agent" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.060542 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f41fef51-25cf-4306-87fa-4f203d7d18e9" containerName="ceilometer-central-agent" Mar 17 09:33:20 crc kubenswrapper[4813]: E0317 09:33:20.060557 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f41fef51-25cf-4306-87fa-4f203d7d18e9" containerName="proxy-httpd" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.060564 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f41fef51-25cf-4306-87fa-4f203d7d18e9" containerName="proxy-httpd" Mar 17 09:33:20 crc kubenswrapper[4813]: E0317 09:33:20.060573 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f41fef51-25cf-4306-87fa-4f203d7d18e9" containerName="sg-core" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.060578 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f41fef51-25cf-4306-87fa-4f203d7d18e9" containerName="sg-core" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.060794 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f41fef51-25cf-4306-87fa-4f203d7d18e9" containerName="ceilometer-central-agent" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.060809 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f41fef51-25cf-4306-87fa-4f203d7d18e9" containerName="proxy-httpd" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.060825 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f41fef51-25cf-4306-87fa-4f203d7d18e9" containerName="sg-core" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.060839 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f41fef51-25cf-4306-87fa-4f203d7d18e9" containerName="ceilometer-notification-agent" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.062351 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.063094 4813 scope.go:117] "RemoveContainer" containerID="01eefccb4733f564717eef9cd02109e444797b79107d8c82d5c9cdf662bdbe39" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.066102 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.066246 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.066736 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.074299 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.107336 4813 scope.go:117] "RemoveContainer" containerID="82687c8db83081af109ede7b426e5712fbcae4c4e3a3097ca211e27cbf1fe77e" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.133709 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.134032 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.134336 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98d79382-0d20-4868-b265-cbf83126f43d-run-httpd\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.134396 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-scripts\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.134484 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98d79382-0d20-4868-b265-cbf83126f43d-log-httpd\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.135245 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-config-data\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.135360 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jcw2z\" (UniqueName: \"kubernetes.io/projected/98d79382-0d20-4868-b265-cbf83126f43d-kube-api-access-jcw2z\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.135460 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.139772 4813 scope.go:117] "RemoveContainer" containerID="a6b6a160c064ec98158a0a506fb892a51ee338a0fbf0fe48426cb9499e167a73" Mar 17 09:33:20 crc kubenswrapper[4813]: E0317 09:33:20.140235 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6b6a160c064ec98158a0a506fb892a51ee338a0fbf0fe48426cb9499e167a73\": container with ID starting with a6b6a160c064ec98158a0a506fb892a51ee338a0fbf0fe48426cb9499e167a73 not found: ID does not exist" containerID="a6b6a160c064ec98158a0a506fb892a51ee338a0fbf0fe48426cb9499e167a73" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.140281 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6b6a160c064ec98158a0a506fb892a51ee338a0fbf0fe48426cb9499e167a73"} err="failed to get container status \"a6b6a160c064ec98158a0a506fb892a51ee338a0fbf0fe48426cb9499e167a73\": rpc error: code = NotFound desc = could not find container \"a6b6a160c064ec98158a0a506fb892a51ee338a0fbf0fe48426cb9499e167a73\": container with ID starting with a6b6a160c064ec98158a0a506fb892a51ee338a0fbf0fe48426cb9499e167a73 not found: ID does not exist" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.140314 4813 scope.go:117] "RemoveContainer" containerID="0533d88c48b4cf268e85d8b71e73c298620f3f3e352445936593f80fb3c9954d" Mar 17 09:33:20 crc kubenswrapper[4813]: E0317 09:33:20.144328 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0533d88c48b4cf268e85d8b71e73c298620f3f3e352445936593f80fb3c9954d\": container with ID starting with 0533d88c48b4cf268e85d8b71e73c298620f3f3e352445936593f80fb3c9954d not found: ID does not exist" containerID="0533d88c48b4cf268e85d8b71e73c298620f3f3e352445936593f80fb3c9954d" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.144660 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0533d88c48b4cf268e85d8b71e73c298620f3f3e352445936593f80fb3c9954d"} err="failed to get container status \"0533d88c48b4cf268e85d8b71e73c298620f3f3e352445936593f80fb3c9954d\": rpc error: code = NotFound desc = could not find container \"0533d88c48b4cf268e85d8b71e73c298620f3f3e352445936593f80fb3c9954d\": container with ID starting with 0533d88c48b4cf268e85d8b71e73c298620f3f3e352445936593f80fb3c9954d not found: ID does not exist" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.144734 4813 scope.go:117] "RemoveContainer" containerID="01eefccb4733f564717eef9cd02109e444797b79107d8c82d5c9cdf662bdbe39" Mar 17 09:33:20 crc kubenswrapper[4813]: E0317 09:33:20.145750 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"01eefccb4733f564717eef9cd02109e444797b79107d8c82d5c9cdf662bdbe39\": container with ID starting with 01eefccb4733f564717eef9cd02109e444797b79107d8c82d5c9cdf662bdbe39 not found: ID does not exist" containerID="01eefccb4733f564717eef9cd02109e444797b79107d8c82d5c9cdf662bdbe39" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.145810 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"01eefccb4733f564717eef9cd02109e444797b79107d8c82d5c9cdf662bdbe39"} err="failed to get container status \"01eefccb4733f564717eef9cd02109e444797b79107d8c82d5c9cdf662bdbe39\": rpc error: code = NotFound desc = could not find container \"01eefccb4733f564717eef9cd02109e444797b79107d8c82d5c9cdf662bdbe39\": container with ID starting with 01eefccb4733f564717eef9cd02109e444797b79107d8c82d5c9cdf662bdbe39 not found: ID does not exist" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.145850 4813 scope.go:117] "RemoveContainer" containerID="82687c8db83081af109ede7b426e5712fbcae4c4e3a3097ca211e27cbf1fe77e" Mar 17 09:33:20 crc kubenswrapper[4813]: E0317 09:33:20.146309 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82687c8db83081af109ede7b426e5712fbcae4c4e3a3097ca211e27cbf1fe77e\": container with ID starting with 82687c8db83081af109ede7b426e5712fbcae4c4e3a3097ca211e27cbf1fe77e not found: ID does not exist" containerID="82687c8db83081af109ede7b426e5712fbcae4c4e3a3097ca211e27cbf1fe77e" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.146338 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82687c8db83081af109ede7b426e5712fbcae4c4e3a3097ca211e27cbf1fe77e"} err="failed to get container status \"82687c8db83081af109ede7b426e5712fbcae4c4e3a3097ca211e27cbf1fe77e\": rpc error: code = NotFound desc = could not find container \"82687c8db83081af109ede7b426e5712fbcae4c4e3a3097ca211e27cbf1fe77e\": container with ID starting with 82687c8db83081af109ede7b426e5712fbcae4c4e3a3097ca211e27cbf1fe77e not found: ID does not exist" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.169915 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.236981 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-config-data\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.237031 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jcw2z\" (UniqueName: \"kubernetes.io/projected/98d79382-0d20-4868-b265-cbf83126f43d-kube-api-access-jcw2z\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.237072 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.237129 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.237168 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.237190 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98d79382-0d20-4868-b265-cbf83126f43d-run-httpd\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.237205 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-scripts\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.237226 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98d79382-0d20-4868-b265-cbf83126f43d-log-httpd\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.237637 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98d79382-0d20-4868-b265-cbf83126f43d-log-httpd\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.239582 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98d79382-0d20-4868-b265-cbf83126f43d-run-httpd\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.241083 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.241525 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.242490 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-config-data\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.243162 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.243944 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-scripts\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: E0317 09:33:20.250831 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf41fef51_25cf_4306_87fa_4f203d7d18e9.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf41fef51_25cf_4306_87fa_4f203d7d18e9.slice/crio-938da892e8016b9e2fd28ea9a2f6d04804e63fdd85adaa92b26e1c84b1116673\": RecentStats: unable to find data in memory cache]" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.261477 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jcw2z\" (UniqueName: \"kubernetes.io/projected/98d79382-0d20-4868-b265-cbf83126f43d-kube-api-access-jcw2z\") pod \"ceilometer-0\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.388641 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.743046 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="baca76f9-d625-4532-9315-8be77971153b" path="/var/lib/kubelet/pods/baca76f9-d625-4532-9315-8be77971153b/volumes" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.744088 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f41fef51-25cf-4306-87fa-4f203d7d18e9" path="/var/lib/kubelet/pods/f41fef51-25cf-4306-87fa-4f203d7d18e9/volumes" Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.897293 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:33:20 crc kubenswrapper[4813]: I0317 09:33:20.942193 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.025145 4813 generic.go:334] "Generic (PLEG): container finished" podID="d318c13f-5f03-40af-86a1-4fd58b6f6b70" containerID="997adf8213b3dd1c07c11b89cc0069428f4f4ee873ac17011cab5fe52d0b58a7" exitCode=0 Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.025227 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d318c13f-5f03-40af-86a1-4fd58b6f6b70","Type":"ContainerDied","Data":"997adf8213b3dd1c07c11b89cc0069428f4f4ee873ac17011cab5fe52d0b58a7"} Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.025293 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"d318c13f-5f03-40af-86a1-4fd58b6f6b70","Type":"ContainerDied","Data":"412a220eea0d0cfb004a96cdb0560254828abb5865076380779b6ca448167f7a"} Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.025317 4813 scope.go:117] "RemoveContainer" containerID="997adf8213b3dd1c07c11b89cc0069428f4f4ee873ac17011cab5fe52d0b58a7" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.025445 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.029378 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98d79382-0d20-4868-b265-cbf83126f43d","Type":"ContainerStarted","Data":"e5009b12f7fea7bf1b476b719c00d644957058dd9e2eecf04a6d6c423a563f5f"} Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.033161 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"69e2e92a-5161-4a09-b59d-464cb4490ff0","Type":"ContainerStarted","Data":"0b9a534ff3bf6cd920494816e2d55e49a081e9a72e266ef15de661be98e603b2"} Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.033194 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"69e2e92a-5161-4a09-b59d-464cb4490ff0","Type":"ContainerStarted","Data":"a5071088567497d0cc96a3626642895fbd9a3c8d81bd904ae0ad1023200d2336"} Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.051449 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d318c13f-5f03-40af-86a1-4fd58b6f6b70-logs\") pod \"d318c13f-5f03-40af-86a1-4fd58b6f6b70\" (UID: \"d318c13f-5f03-40af-86a1-4fd58b6f6b70\") " Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.051619 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d318c13f-5f03-40af-86a1-4fd58b6f6b70-combined-ca-bundle\") pod \"d318c13f-5f03-40af-86a1-4fd58b6f6b70\" (UID: \"d318c13f-5f03-40af-86a1-4fd58b6f6b70\") " Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.051653 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d318c13f-5f03-40af-86a1-4fd58b6f6b70-config-data\") pod \"d318c13f-5f03-40af-86a1-4fd58b6f6b70\" (UID: \"d318c13f-5f03-40af-86a1-4fd58b6f6b70\") " Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.051709 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pd78\" (UniqueName: \"kubernetes.io/projected/d318c13f-5f03-40af-86a1-4fd58b6f6b70-kube-api-access-8pd78\") pod \"d318c13f-5f03-40af-86a1-4fd58b6f6b70\" (UID: \"d318c13f-5f03-40af-86a1-4fd58b6f6b70\") " Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.052030 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d318c13f-5f03-40af-86a1-4fd58b6f6b70-logs" (OuterVolumeSpecName: "logs") pod "d318c13f-5f03-40af-86a1-4fd58b6f6b70" (UID: "d318c13f-5f03-40af-86a1-4fd58b6f6b70"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.057913 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d318c13f-5f03-40af-86a1-4fd58b6f6b70-kube-api-access-8pd78" (OuterVolumeSpecName: "kube-api-access-8pd78") pod "d318c13f-5f03-40af-86a1-4fd58b6f6b70" (UID: "d318c13f-5f03-40af-86a1-4fd58b6f6b70"). InnerVolumeSpecName "kube-api-access-8pd78". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.061581 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.06155863 podStartE2EDuration="2.06155863s" podCreationTimestamp="2026-03-17 09:33:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:33:21.051091152 +0000 UTC m=+1423.151894651" watchObservedRunningTime="2026-03-17 09:33:21.06155863 +0000 UTC m=+1423.162362129" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.062370 4813 scope.go:117] "RemoveContainer" containerID="96b00c2c3649f24a02186e3513367217ea4a0c7995fc302aaf961370b6f222d6" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.079496 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d318c13f-5f03-40af-86a1-4fd58b6f6b70-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d318c13f-5f03-40af-86a1-4fd58b6f6b70" (UID: "d318c13f-5f03-40af-86a1-4fd58b6f6b70"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.084454 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d318c13f-5f03-40af-86a1-4fd58b6f6b70-config-data" (OuterVolumeSpecName: "config-data") pod "d318c13f-5f03-40af-86a1-4fd58b6f6b70" (UID: "d318c13f-5f03-40af-86a1-4fd58b6f6b70"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.090611 4813 scope.go:117] "RemoveContainer" containerID="997adf8213b3dd1c07c11b89cc0069428f4f4ee873ac17011cab5fe52d0b58a7" Mar 17 09:33:21 crc kubenswrapper[4813]: E0317 09:33:21.091230 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"997adf8213b3dd1c07c11b89cc0069428f4f4ee873ac17011cab5fe52d0b58a7\": container with ID starting with 997adf8213b3dd1c07c11b89cc0069428f4f4ee873ac17011cab5fe52d0b58a7 not found: ID does not exist" containerID="997adf8213b3dd1c07c11b89cc0069428f4f4ee873ac17011cab5fe52d0b58a7" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.091269 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"997adf8213b3dd1c07c11b89cc0069428f4f4ee873ac17011cab5fe52d0b58a7"} err="failed to get container status \"997adf8213b3dd1c07c11b89cc0069428f4f4ee873ac17011cab5fe52d0b58a7\": rpc error: code = NotFound desc = could not find container \"997adf8213b3dd1c07c11b89cc0069428f4f4ee873ac17011cab5fe52d0b58a7\": container with ID starting with 997adf8213b3dd1c07c11b89cc0069428f4f4ee873ac17011cab5fe52d0b58a7 not found: ID does not exist" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.091293 4813 scope.go:117] "RemoveContainer" containerID="96b00c2c3649f24a02186e3513367217ea4a0c7995fc302aaf961370b6f222d6" Mar 17 09:33:21 crc kubenswrapper[4813]: E0317 09:33:21.091590 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"96b00c2c3649f24a02186e3513367217ea4a0c7995fc302aaf961370b6f222d6\": container with ID starting with 96b00c2c3649f24a02186e3513367217ea4a0c7995fc302aaf961370b6f222d6 not found: ID does not exist" containerID="96b00c2c3649f24a02186e3513367217ea4a0c7995fc302aaf961370b6f222d6" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.091641 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"96b00c2c3649f24a02186e3513367217ea4a0c7995fc302aaf961370b6f222d6"} err="failed to get container status \"96b00c2c3649f24a02186e3513367217ea4a0c7995fc302aaf961370b6f222d6\": rpc error: code = NotFound desc = could not find container \"96b00c2c3649f24a02186e3513367217ea4a0c7995fc302aaf961370b6f222d6\": container with ID starting with 96b00c2c3649f24a02186e3513367217ea4a0c7995fc302aaf961370b6f222d6 not found: ID does not exist" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.154031 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d318c13f-5f03-40af-86a1-4fd58b6f6b70-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.154070 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d318c13f-5f03-40af-86a1-4fd58b6f6b70-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.154083 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pd78\" (UniqueName: \"kubernetes.io/projected/d318c13f-5f03-40af-86a1-4fd58b6f6b70-kube-api-access-8pd78\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.154096 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d318c13f-5f03-40af-86a1-4fd58b6f6b70-logs\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.366799 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.376741 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.396187 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 17 09:33:21 crc kubenswrapper[4813]: E0317 09:33:21.396534 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d318c13f-5f03-40af-86a1-4fd58b6f6b70" containerName="nova-api-api" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.396549 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d318c13f-5f03-40af-86a1-4fd58b6f6b70" containerName="nova-api-api" Mar 17 09:33:21 crc kubenswrapper[4813]: E0317 09:33:21.396580 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d318c13f-5f03-40af-86a1-4fd58b6f6b70" containerName="nova-api-log" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.396587 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d318c13f-5f03-40af-86a1-4fd58b6f6b70" containerName="nova-api-log" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.396784 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d318c13f-5f03-40af-86a1-4fd58b6f6b70" containerName="nova-api-api" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.396797 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d318c13f-5f03-40af-86a1-4fd58b6f6b70" containerName="nova-api-log" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.397664 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.400912 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.451695 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.459134 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qql8x\" (UniqueName: \"kubernetes.io/projected/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-kube-api-access-qql8x\") pod \"nova-api-0\" (UID: \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\") " pod="openstack/nova-api-0" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.459240 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-config-data\") pod \"nova-api-0\" (UID: \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\") " pod="openstack/nova-api-0" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.459281 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-logs\") pod \"nova-api-0\" (UID: \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\") " pod="openstack/nova-api-0" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.459310 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\") " pod="openstack/nova-api-0" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.560878 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-config-data\") pod \"nova-api-0\" (UID: \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\") " pod="openstack/nova-api-0" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.561027 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-logs\") pod \"nova-api-0\" (UID: \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\") " pod="openstack/nova-api-0" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.561113 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\") " pod="openstack/nova-api-0" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.561299 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qql8x\" (UniqueName: \"kubernetes.io/projected/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-kube-api-access-qql8x\") pod \"nova-api-0\" (UID: \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\") " pod="openstack/nova-api-0" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.563421 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-logs\") pod \"nova-api-0\" (UID: \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\") " pod="openstack/nova-api-0" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.566206 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\") " pod="openstack/nova-api-0" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.567517 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-config-data\") pod \"nova-api-0\" (UID: \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\") " pod="openstack/nova-api-0" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.578313 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qql8x\" (UniqueName: \"kubernetes.io/projected/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-kube-api-access-qql8x\") pod \"nova-api-0\" (UID: \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\") " pod="openstack/nova-api-0" Mar 17 09:33:21 crc kubenswrapper[4813]: I0317 09:33:21.764179 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 09:33:22 crc kubenswrapper[4813]: I0317 09:33:22.044961 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98d79382-0d20-4868-b265-cbf83126f43d","Type":"ContainerStarted","Data":"f6e0137118032ea186557ccee3fa1aaf40e2eef858ffe86dab662cb8c5111fd6"} Mar 17 09:33:22 crc kubenswrapper[4813]: I0317 09:33:22.229572 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 17 09:33:22 crc kubenswrapper[4813]: I0317 09:33:22.744943 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d318c13f-5f03-40af-86a1-4fd58b6f6b70" path="/var/lib/kubelet/pods/d318c13f-5f03-40af-86a1-4fd58b6f6b70/volumes" Mar 17 09:33:23 crc kubenswrapper[4813]: I0317 09:33:23.070780 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98d79382-0d20-4868-b265-cbf83126f43d","Type":"ContainerStarted","Data":"73117d1581b3b0162cf932507df289dd3c6ab947a79ae33d8c35fa36600444e4"} Mar 17 09:33:23 crc kubenswrapper[4813]: I0317 09:33:23.074326 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9e13c70a-9e05-4d99-a662-ac6cff2f3abc","Type":"ContainerStarted","Data":"ee37e13ddff7378536b800c5d4d189cffbc92a4388fc73492ed5f52ff2f7c3a4"} Mar 17 09:33:23 crc kubenswrapper[4813]: I0317 09:33:23.074372 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9e13c70a-9e05-4d99-a662-ac6cff2f3abc","Type":"ContainerStarted","Data":"1b05ed8d644fe11cf3ceb7f433f9d73358cfe9ad0832c1b180007725a6d5d9ea"} Mar 17 09:33:23 crc kubenswrapper[4813]: I0317 09:33:23.074386 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9e13c70a-9e05-4d99-a662-ac6cff2f3abc","Type":"ContainerStarted","Data":"2c01de7864b85f0fddca34ef67a17da121788f3745742569446f04cefbbeaa08"} Mar 17 09:33:23 crc kubenswrapper[4813]: I0317 09:33:23.095844 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.095822802 podStartE2EDuration="2.095822802s" podCreationTimestamp="2026-03-17 09:33:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:33:23.090893767 +0000 UTC m=+1425.191697276" watchObservedRunningTime="2026-03-17 09:33:23.095822802 +0000 UTC m=+1425.196626301" Mar 17 09:33:23 crc kubenswrapper[4813]: I0317 09:33:23.362851 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Mar 17 09:33:24 crc kubenswrapper[4813]: I0317 09:33:24.087441 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98d79382-0d20-4868-b265-cbf83126f43d","Type":"ContainerStarted","Data":"231fb631697d176a2411dfed49a192eac36fc630290640149cc808125041f064"} Mar 17 09:33:24 crc kubenswrapper[4813]: I0317 09:33:24.742941 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 17 09:33:25 crc kubenswrapper[4813]: I0317 09:33:25.328214 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 17 09:33:25 crc kubenswrapper[4813]: I0317 09:33:25.328549 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 17 09:33:26 crc kubenswrapper[4813]: I0317 09:33:26.119356 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98d79382-0d20-4868-b265-cbf83126f43d","Type":"ContainerStarted","Data":"90928cf06200bd034b239dd4beb66909b1ef8722ac8df18ff012dfba7270dec0"} Mar 17 09:33:26 crc kubenswrapper[4813]: I0317 09:33:26.119675 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 17 09:33:26 crc kubenswrapper[4813]: I0317 09:33:26.154320 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.205268618 podStartE2EDuration="6.154266141s" podCreationTimestamp="2026-03-17 09:33:20 +0000 UTC" firstStartedPulling="2026-03-17 09:33:20.891361612 +0000 UTC m=+1422.992165111" lastFinishedPulling="2026-03-17 09:33:24.840359115 +0000 UTC m=+1426.941162634" observedRunningTime="2026-03-17 09:33:26.145413753 +0000 UTC m=+1428.246217262" watchObservedRunningTime="2026-03-17 09:33:26.154266141 +0000 UTC m=+1428.255069650" Mar 17 09:33:26 crc kubenswrapper[4813]: I0317 09:33:26.341001 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4b69ffd2-16ba-45af-b3c9-d4f8b3effebe" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 09:33:26 crc kubenswrapper[4813]: I0317 09:33:26.341878 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="4b69ffd2-16ba-45af-b3c9-d4f8b3effebe" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.203:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 09:33:26 crc kubenswrapper[4813]: I0317 09:33:26.681854 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Mar 17 09:33:29 crc kubenswrapper[4813]: I0317 09:33:29.730961 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 17 09:33:29 crc kubenswrapper[4813]: I0317 09:33:29.764510 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 17 09:33:30 crc kubenswrapper[4813]: I0317 09:33:30.183990 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 17 09:33:31 crc kubenswrapper[4813]: I0317 09:33:31.764837 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 17 09:33:31 crc kubenswrapper[4813]: I0317 09:33:31.764895 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 17 09:33:32 crc kubenswrapper[4813]: I0317 09:33:32.806753 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9e13c70a-9e05-4d99-a662-ac6cff2f3abc" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.207:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 09:33:32 crc kubenswrapper[4813]: I0317 09:33:32.806760 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9e13c70a-9e05-4d99-a662-ac6cff2f3abc" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.207:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 09:33:33 crc kubenswrapper[4813]: I0317 09:33:33.327817 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 17 09:33:33 crc kubenswrapper[4813]: I0317 09:33:33.327899 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 17 09:33:35 crc kubenswrapper[4813]: I0317 09:33:35.355955 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 17 09:33:35 crc kubenswrapper[4813]: I0317 09:33:35.361694 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 17 09:33:35 crc kubenswrapper[4813]: I0317 09:33:35.363714 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 17 09:33:36 crc kubenswrapper[4813]: I0317 09:33:36.249744 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 17 09:33:38 crc kubenswrapper[4813]: I0317 09:33:38.267453 4813 generic.go:334] "Generic (PLEG): container finished" podID="845fdf7b-a3b0-487b-9bcd-efcc3b2b8513" containerID="987f42380506732cb7bccd1f21a56d81e217d17ebf5e5f8bb56bd3397368fdac" exitCode=137 Mar 17 09:33:38 crc kubenswrapper[4813]: I0317 09:33:38.267577 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"845fdf7b-a3b0-487b-9bcd-efcc3b2b8513","Type":"ContainerDied","Data":"987f42380506732cb7bccd1f21a56d81e217d17ebf5e5f8bb56bd3397368fdac"} Mar 17 09:33:38 crc kubenswrapper[4813]: I0317 09:33:38.268169 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"845fdf7b-a3b0-487b-9bcd-efcc3b2b8513","Type":"ContainerDied","Data":"e7013bd9fd2db2fb53e84999c9ed74477919c76cacc8da0eec310d8b91807f8b"} Mar 17 09:33:38 crc kubenswrapper[4813]: I0317 09:33:38.268196 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7013bd9fd2db2fb53e84999c9ed74477919c76cacc8da0eec310d8b91807f8b" Mar 17 09:33:38 crc kubenswrapper[4813]: I0317 09:33:38.323526 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:38 crc kubenswrapper[4813]: I0317 09:33:38.510979 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lsj52\" (UniqueName: \"kubernetes.io/projected/845fdf7b-a3b0-487b-9bcd-efcc3b2b8513-kube-api-access-lsj52\") pod \"845fdf7b-a3b0-487b-9bcd-efcc3b2b8513\" (UID: \"845fdf7b-a3b0-487b-9bcd-efcc3b2b8513\") " Mar 17 09:33:38 crc kubenswrapper[4813]: I0317 09:33:38.511119 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/845fdf7b-a3b0-487b-9bcd-efcc3b2b8513-combined-ca-bundle\") pod \"845fdf7b-a3b0-487b-9bcd-efcc3b2b8513\" (UID: \"845fdf7b-a3b0-487b-9bcd-efcc3b2b8513\") " Mar 17 09:33:38 crc kubenswrapper[4813]: I0317 09:33:38.511308 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/845fdf7b-a3b0-487b-9bcd-efcc3b2b8513-config-data\") pod \"845fdf7b-a3b0-487b-9bcd-efcc3b2b8513\" (UID: \"845fdf7b-a3b0-487b-9bcd-efcc3b2b8513\") " Mar 17 09:33:38 crc kubenswrapper[4813]: I0317 09:33:38.524105 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/845fdf7b-a3b0-487b-9bcd-efcc3b2b8513-kube-api-access-lsj52" (OuterVolumeSpecName: "kube-api-access-lsj52") pod "845fdf7b-a3b0-487b-9bcd-efcc3b2b8513" (UID: "845fdf7b-a3b0-487b-9bcd-efcc3b2b8513"). InnerVolumeSpecName "kube-api-access-lsj52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:33:38 crc kubenswrapper[4813]: I0317 09:33:38.545901 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/845fdf7b-a3b0-487b-9bcd-efcc3b2b8513-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "845fdf7b-a3b0-487b-9bcd-efcc3b2b8513" (UID: "845fdf7b-a3b0-487b-9bcd-efcc3b2b8513"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:38 crc kubenswrapper[4813]: I0317 09:33:38.567048 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/845fdf7b-a3b0-487b-9bcd-efcc3b2b8513-config-data" (OuterVolumeSpecName: "config-data") pod "845fdf7b-a3b0-487b-9bcd-efcc3b2b8513" (UID: "845fdf7b-a3b0-487b-9bcd-efcc3b2b8513"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:38 crc kubenswrapper[4813]: I0317 09:33:38.614126 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/845fdf7b-a3b0-487b-9bcd-efcc3b2b8513-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:38 crc kubenswrapper[4813]: I0317 09:33:38.614165 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lsj52\" (UniqueName: \"kubernetes.io/projected/845fdf7b-a3b0-487b-9bcd-efcc3b2b8513-kube-api-access-lsj52\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:38 crc kubenswrapper[4813]: I0317 09:33:38.614179 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/845fdf7b-a3b0-487b-9bcd-efcc3b2b8513-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.279243 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.310562 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.319366 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.356475 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 17 09:33:39 crc kubenswrapper[4813]: E0317 09:33:39.357433 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="845fdf7b-a3b0-487b-9bcd-efcc3b2b8513" containerName="nova-cell1-novncproxy-novncproxy" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.357530 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="845fdf7b-a3b0-487b-9bcd-efcc3b2b8513" containerName="nova-cell1-novncproxy-novncproxy" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.358066 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="845fdf7b-a3b0-487b-9bcd-efcc3b2b8513" containerName="nova-cell1-novncproxy-novncproxy" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.359472 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.369716 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.369940 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.370217 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.370360 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.431116 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cc3a926-a367-45b1-9d38-6f673720e71e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9cc3a926-a367-45b1-9d38-6f673720e71e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.431228 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cc3a926-a367-45b1-9d38-6f673720e71e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9cc3a926-a367-45b1-9d38-6f673720e71e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.431313 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cc3a926-a367-45b1-9d38-6f673720e71e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9cc3a926-a367-45b1-9d38-6f673720e71e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.431356 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pjmp9\" (UniqueName: \"kubernetes.io/projected/9cc3a926-a367-45b1-9d38-6f673720e71e-kube-api-access-pjmp9\") pod \"nova-cell1-novncproxy-0\" (UID: \"9cc3a926-a367-45b1-9d38-6f673720e71e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.431507 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cc3a926-a367-45b1-9d38-6f673720e71e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9cc3a926-a367-45b1-9d38-6f673720e71e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.533996 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cc3a926-a367-45b1-9d38-6f673720e71e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9cc3a926-a367-45b1-9d38-6f673720e71e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.534063 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cc3a926-a367-45b1-9d38-6f673720e71e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9cc3a926-a367-45b1-9d38-6f673720e71e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.534133 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cc3a926-a367-45b1-9d38-6f673720e71e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9cc3a926-a367-45b1-9d38-6f673720e71e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.534165 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pjmp9\" (UniqueName: \"kubernetes.io/projected/9cc3a926-a367-45b1-9d38-6f673720e71e-kube-api-access-pjmp9\") pod \"nova-cell1-novncproxy-0\" (UID: \"9cc3a926-a367-45b1-9d38-6f673720e71e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.534209 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cc3a926-a367-45b1-9d38-6f673720e71e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9cc3a926-a367-45b1-9d38-6f673720e71e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.539542 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cc3a926-a367-45b1-9d38-6f673720e71e-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9cc3a926-a367-45b1-9d38-6f673720e71e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.539631 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cc3a926-a367-45b1-9d38-6f673720e71e-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"9cc3a926-a367-45b1-9d38-6f673720e71e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.540411 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/9cc3a926-a367-45b1-9d38-6f673720e71e-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"9cc3a926-a367-45b1-9d38-6f673720e71e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.541300 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cc3a926-a367-45b1-9d38-6f673720e71e-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"9cc3a926-a367-45b1-9d38-6f673720e71e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.563416 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pjmp9\" (UniqueName: \"kubernetes.io/projected/9cc3a926-a367-45b1-9d38-6f673720e71e-kube-api-access-pjmp9\") pod \"nova-cell1-novncproxy-0\" (UID: \"9cc3a926-a367-45b1-9d38-6f673720e71e\") " pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.701940 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.764647 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 17 09:33:39 crc kubenswrapper[4813]: I0317 09:33:39.764752 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 17 09:33:40 crc kubenswrapper[4813]: I0317 09:33:40.201102 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Mar 17 09:33:40 crc kubenswrapper[4813]: I0317 09:33:40.290263 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9cc3a926-a367-45b1-9d38-6f673720e71e","Type":"ContainerStarted","Data":"02ceee19963383ac243d688e9eaef18defbab8202c87319c8a4d0cbbf05bd013"} Mar 17 09:33:40 crc kubenswrapper[4813]: E0317 09:33:40.725109 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod845fdf7b_a3b0_487b_9bcd_efcc3b2b8513.slice\": RecentStats: unable to find data in memory cache]" Mar 17 09:33:40 crc kubenswrapper[4813]: I0317 09:33:40.751428 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="845fdf7b-a3b0-487b-9bcd-efcc3b2b8513" path="/var/lib/kubelet/pods/845fdf7b-a3b0-487b-9bcd-efcc3b2b8513/volumes" Mar 17 09:33:41 crc kubenswrapper[4813]: I0317 09:33:41.305270 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"9cc3a926-a367-45b1-9d38-6f673720e71e","Type":"ContainerStarted","Data":"8eaf342a911d30ae3365d03d3fb18e9b893d974a220eee921359686d9afee53d"} Mar 17 09:33:41 crc kubenswrapper[4813]: I0317 09:33:41.340680 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.340649574 podStartE2EDuration="2.340649574s" podCreationTimestamp="2026-03-17 09:33:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:33:41.329091421 +0000 UTC m=+1443.429894950" watchObservedRunningTime="2026-03-17 09:33:41.340649574 +0000 UTC m=+1443.441453113" Mar 17 09:33:41 crc kubenswrapper[4813]: I0317 09:33:41.771294 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 17 09:33:41 crc kubenswrapper[4813]: I0317 09:33:41.771380 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 17 09:33:41 crc kubenswrapper[4813]: I0317 09:33:41.776826 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 17 09:33:41 crc kubenswrapper[4813]: I0317 09:33:41.778118 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.025167 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-cv6cc"] Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.026737 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.047763 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-cv6cc"] Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.094307 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-cv6cc\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.094475 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cv6nf\" (UniqueName: \"kubernetes.io/projected/0b1a228b-5bd8-4340-a286-cf609851c75d-kube-api-access-cv6nf\") pod \"dnsmasq-dns-59cf4bdb65-cv6cc\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.094673 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-config\") pod \"dnsmasq-dns-59cf4bdb65-cv6cc\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.097225 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-cv6cc\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.097254 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-cv6cc\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.097274 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-cv6cc\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.198076 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-config\") pod \"dnsmasq-dns-59cf4bdb65-cv6cc\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.198115 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-cv6cc\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.198131 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-cv6cc\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.198149 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-cv6cc\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.199090 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-cv6cc\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.199141 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cv6nf\" (UniqueName: \"kubernetes.io/projected/0b1a228b-5bd8-4340-a286-cf609851c75d-kube-api-access-cv6nf\") pod \"dnsmasq-dns-59cf4bdb65-cv6cc\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.198987 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-dns-svc\") pod \"dnsmasq-dns-59cf4bdb65-cv6cc\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.199088 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-dns-swift-storage-0\") pod \"dnsmasq-dns-59cf4bdb65-cv6cc\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.198987 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-config\") pod \"dnsmasq-dns-59cf4bdb65-cv6cc\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.198991 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-ovsdbserver-sb\") pod \"dnsmasq-dns-59cf4bdb65-cv6cc\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.199585 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-ovsdbserver-nb\") pod \"dnsmasq-dns-59cf4bdb65-cv6cc\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.219236 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cv6nf\" (UniqueName: \"kubernetes.io/projected/0b1a228b-5bd8-4340-a286-cf609851c75d-kube-api-access-cv6nf\") pod \"dnsmasq-dns-59cf4bdb65-cv6cc\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.346356 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:42 crc kubenswrapper[4813]: I0317 09:33:42.830042 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-cv6cc"] Mar 17 09:33:43 crc kubenswrapper[4813]: I0317 09:33:43.326015 4813 generic.go:334] "Generic (PLEG): container finished" podID="0b1a228b-5bd8-4340-a286-cf609851c75d" containerID="bc6ab5b77103f301ace06b4bf22d90de46dbfb3bd748a19e1774b5d8eceb994d" exitCode=0 Mar 17 09:33:43 crc kubenswrapper[4813]: I0317 09:33:43.326574 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" event={"ID":"0b1a228b-5bd8-4340-a286-cf609851c75d","Type":"ContainerDied","Data":"bc6ab5b77103f301ace06b4bf22d90de46dbfb3bd748a19e1774b5d8eceb994d"} Mar 17 09:33:43 crc kubenswrapper[4813]: I0317 09:33:43.326636 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" event={"ID":"0b1a228b-5bd8-4340-a286-cf609851c75d","Type":"ContainerStarted","Data":"e2570020dc39abf30bcb2883a46a66b3ee8e5e4dece7b1df0d986e54f24b5437"} Mar 17 09:33:43 crc kubenswrapper[4813]: I0317 09:33:43.854097 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:33:43 crc kubenswrapper[4813]: I0317 09:33:43.854632 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="98d79382-0d20-4868-b265-cbf83126f43d" containerName="ceilometer-central-agent" containerID="cri-o://f6e0137118032ea186557ccee3fa1aaf40e2eef858ffe86dab662cb8c5111fd6" gracePeriod=30 Mar 17 09:33:43 crc kubenswrapper[4813]: I0317 09:33:43.854750 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="98d79382-0d20-4868-b265-cbf83126f43d" containerName="ceilometer-notification-agent" containerID="cri-o://73117d1581b3b0162cf932507df289dd3c6ab947a79ae33d8c35fa36600444e4" gracePeriod=30 Mar 17 09:33:43 crc kubenswrapper[4813]: I0317 09:33:43.854757 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="98d79382-0d20-4868-b265-cbf83126f43d" containerName="sg-core" containerID="cri-o://231fb631697d176a2411dfed49a192eac36fc630290640149cc808125041f064" gracePeriod=30 Mar 17 09:33:43 crc kubenswrapper[4813]: I0317 09:33:43.854926 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="98d79382-0d20-4868-b265-cbf83126f43d" containerName="proxy-httpd" containerID="cri-o://90928cf06200bd034b239dd4beb66909b1ef8722ac8df18ff012dfba7270dec0" gracePeriod=30 Mar 17 09:33:43 crc kubenswrapper[4813]: I0317 09:33:43.882382 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="98d79382-0d20-4868-b265-cbf83126f43d" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.114058 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.114113 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.114154 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.114888 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4ba33e308a10bfb001c8289b81c4de34337bf84d52f75bd8b871b725edd0f79b"} pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.114943 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" containerID="cri-o://4ba33e308a10bfb001c8289b81c4de34337bf84d52f75bd8b871b725edd0f79b" gracePeriod=600 Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.213647 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.343323 4813 generic.go:334] "Generic (PLEG): container finished" podID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerID="4ba33e308a10bfb001c8289b81c4de34337bf84d52f75bd8b871b725edd0f79b" exitCode=0 Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.343772 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerDied","Data":"4ba33e308a10bfb001c8289b81c4de34337bf84d52f75bd8b871b725edd0f79b"} Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.343805 4813 scope.go:117] "RemoveContainer" containerID="d74d47d3497863468b9c4543e558eb883310048990c8823420967f1910298e01" Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.350438 4813 generic.go:334] "Generic (PLEG): container finished" podID="98d79382-0d20-4868-b265-cbf83126f43d" containerID="90928cf06200bd034b239dd4beb66909b1ef8722ac8df18ff012dfba7270dec0" exitCode=0 Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.350469 4813 generic.go:334] "Generic (PLEG): container finished" podID="98d79382-0d20-4868-b265-cbf83126f43d" containerID="231fb631697d176a2411dfed49a192eac36fc630290640149cc808125041f064" exitCode=2 Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.350481 4813 generic.go:334] "Generic (PLEG): container finished" podID="98d79382-0d20-4868-b265-cbf83126f43d" containerID="f6e0137118032ea186557ccee3fa1aaf40e2eef858ffe86dab662cb8c5111fd6" exitCode=0 Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.350519 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98d79382-0d20-4868-b265-cbf83126f43d","Type":"ContainerDied","Data":"90928cf06200bd034b239dd4beb66909b1ef8722ac8df18ff012dfba7270dec0"} Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.350541 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98d79382-0d20-4868-b265-cbf83126f43d","Type":"ContainerDied","Data":"231fb631697d176a2411dfed49a192eac36fc630290640149cc808125041f064"} Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.350555 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98d79382-0d20-4868-b265-cbf83126f43d","Type":"ContainerDied","Data":"f6e0137118032ea186557ccee3fa1aaf40e2eef858ffe86dab662cb8c5111fd6"} Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.353527 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" event={"ID":"0b1a228b-5bd8-4340-a286-cf609851c75d","Type":"ContainerStarted","Data":"c1a329c4c5592bc62a8a342e76504c2ee24fde54a809235d3f5d2308031e86bf"} Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.353591 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9e13c70a-9e05-4d99-a662-ac6cff2f3abc" containerName="nova-api-log" containerID="cri-o://1b05ed8d644fe11cf3ceb7f433f9d73358cfe9ad0832c1b180007725a6d5d9ea" gracePeriod=30 Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.353693 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9e13c70a-9e05-4d99-a662-ac6cff2f3abc" containerName="nova-api-api" containerID="cri-o://ee37e13ddff7378536b800c5d4d189cffbc92a4388fc73492ed5f52ff2f7c3a4" gracePeriod=30 Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.353914 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.391812 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" podStartSLOduration=3.391791833 podStartE2EDuration="3.391791833s" podCreationTimestamp="2026-03-17 09:33:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:33:44.38148446 +0000 UTC m=+1446.482287959" watchObservedRunningTime="2026-03-17 09:33:44.391791833 +0000 UTC m=+1446.492595332" Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.704519 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.772852 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.950401 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-scripts\") pod \"98d79382-0d20-4868-b265-cbf83126f43d\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.950501 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98d79382-0d20-4868-b265-cbf83126f43d-log-httpd\") pod \"98d79382-0d20-4868-b265-cbf83126f43d\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.950548 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-ceilometer-tls-certs\") pod \"98d79382-0d20-4868-b265-cbf83126f43d\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.950627 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-combined-ca-bundle\") pod \"98d79382-0d20-4868-b265-cbf83126f43d\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.950658 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-config-data\") pod \"98d79382-0d20-4868-b265-cbf83126f43d\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.950711 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-sg-core-conf-yaml\") pod \"98d79382-0d20-4868-b265-cbf83126f43d\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.950735 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98d79382-0d20-4868-b265-cbf83126f43d-run-httpd\") pod \"98d79382-0d20-4868-b265-cbf83126f43d\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.950793 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jcw2z\" (UniqueName: \"kubernetes.io/projected/98d79382-0d20-4868-b265-cbf83126f43d-kube-api-access-jcw2z\") pod \"98d79382-0d20-4868-b265-cbf83126f43d\" (UID: \"98d79382-0d20-4868-b265-cbf83126f43d\") " Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.952364 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98d79382-0d20-4868-b265-cbf83126f43d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "98d79382-0d20-4868-b265-cbf83126f43d" (UID: "98d79382-0d20-4868-b265-cbf83126f43d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.953029 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/98d79382-0d20-4868-b265-cbf83126f43d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "98d79382-0d20-4868-b265-cbf83126f43d" (UID: "98d79382-0d20-4868-b265-cbf83126f43d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.956051 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98d79382-0d20-4868-b265-cbf83126f43d-kube-api-access-jcw2z" (OuterVolumeSpecName: "kube-api-access-jcw2z") pod "98d79382-0d20-4868-b265-cbf83126f43d" (UID: "98d79382-0d20-4868-b265-cbf83126f43d"). InnerVolumeSpecName "kube-api-access-jcw2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.957695 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-scripts" (OuterVolumeSpecName: "scripts") pod "98d79382-0d20-4868-b265-cbf83126f43d" (UID: "98d79382-0d20-4868-b265-cbf83126f43d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:44 crc kubenswrapper[4813]: I0317 09:33:44.985078 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "98d79382-0d20-4868-b265-cbf83126f43d" (UID: "98d79382-0d20-4868-b265-cbf83126f43d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.021439 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "98d79382-0d20-4868-b265-cbf83126f43d" (UID: "98d79382-0d20-4868-b265-cbf83126f43d"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.053223 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jcw2z\" (UniqueName: \"kubernetes.io/projected/98d79382-0d20-4868-b265-cbf83126f43d-kube-api-access-jcw2z\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.053528 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.053542 4813 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98d79382-0d20-4868-b265-cbf83126f43d-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.053552 4813 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.053560 4813 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.053571 4813 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/98d79382-0d20-4868-b265-cbf83126f43d-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.057832 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "98d79382-0d20-4868-b265-cbf83126f43d" (UID: "98d79382-0d20-4868-b265-cbf83126f43d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.071078 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-config-data" (OuterVolumeSpecName: "config-data") pod "98d79382-0d20-4868-b265-cbf83126f43d" (UID: "98d79382-0d20-4868-b265-cbf83126f43d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.154692 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.155345 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98d79382-0d20-4868-b265-cbf83126f43d-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.364550 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerStarted","Data":"cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368"} Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.367214 4813 generic.go:334] "Generic (PLEG): container finished" podID="98d79382-0d20-4868-b265-cbf83126f43d" containerID="73117d1581b3b0162cf932507df289dd3c6ab947a79ae33d8c35fa36600444e4" exitCode=0 Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.367304 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98d79382-0d20-4868-b265-cbf83126f43d","Type":"ContainerDied","Data":"73117d1581b3b0162cf932507df289dd3c6ab947a79ae33d8c35fa36600444e4"} Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.367343 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"98d79382-0d20-4868-b265-cbf83126f43d","Type":"ContainerDied","Data":"e5009b12f7fea7bf1b476b719c00d644957058dd9e2eecf04a6d6c423a563f5f"} Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.367371 4813 scope.go:117] "RemoveContainer" containerID="90928cf06200bd034b239dd4beb66909b1ef8722ac8df18ff012dfba7270dec0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.367534 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.369545 4813 generic.go:334] "Generic (PLEG): container finished" podID="9e13c70a-9e05-4d99-a662-ac6cff2f3abc" containerID="1b05ed8d644fe11cf3ceb7f433f9d73358cfe9ad0832c1b180007725a6d5d9ea" exitCode=143 Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.369748 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9e13c70a-9e05-4d99-a662-ac6cff2f3abc","Type":"ContainerDied","Data":"1b05ed8d644fe11cf3ceb7f433f9d73358cfe9ad0832c1b180007725a6d5d9ea"} Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.388682 4813 scope.go:117] "RemoveContainer" containerID="231fb631697d176a2411dfed49a192eac36fc630290640149cc808125041f064" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.409861 4813 scope.go:117] "RemoveContainer" containerID="73117d1581b3b0162cf932507df289dd3c6ab947a79ae33d8c35fa36600444e4" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.424646 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.441175 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.443794 4813 scope.go:117] "RemoveContainer" containerID="f6e0137118032ea186557ccee3fa1aaf40e2eef858ffe86dab662cb8c5111fd6" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.450659 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:33:45 crc kubenswrapper[4813]: E0317 09:33:45.451105 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98d79382-0d20-4868-b265-cbf83126f43d" containerName="proxy-httpd" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.451124 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="98d79382-0d20-4868-b265-cbf83126f43d" containerName="proxy-httpd" Mar 17 09:33:45 crc kubenswrapper[4813]: E0317 09:33:45.451145 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98d79382-0d20-4868-b265-cbf83126f43d" containerName="ceilometer-central-agent" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.451154 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="98d79382-0d20-4868-b265-cbf83126f43d" containerName="ceilometer-central-agent" Mar 17 09:33:45 crc kubenswrapper[4813]: E0317 09:33:45.451168 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98d79382-0d20-4868-b265-cbf83126f43d" containerName="sg-core" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.451174 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="98d79382-0d20-4868-b265-cbf83126f43d" containerName="sg-core" Mar 17 09:33:45 crc kubenswrapper[4813]: E0317 09:33:45.451188 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98d79382-0d20-4868-b265-cbf83126f43d" containerName="ceilometer-notification-agent" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.451197 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="98d79382-0d20-4868-b265-cbf83126f43d" containerName="ceilometer-notification-agent" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.451366 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="98d79382-0d20-4868-b265-cbf83126f43d" containerName="ceilometer-notification-agent" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.451382 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="98d79382-0d20-4868-b265-cbf83126f43d" containerName="sg-core" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.451389 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="98d79382-0d20-4868-b265-cbf83126f43d" containerName="ceilometer-central-agent" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.451397 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="98d79382-0d20-4868-b265-cbf83126f43d" containerName="proxy-httpd" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.452978 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.454823 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.457101 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.457345 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.460746 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.461060 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.461106 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.461233 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-run-httpd\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.461264 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-config-data\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.461362 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.461392 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-scripts\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.461427 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-log-httpd\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.461498 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hdk8\" (UniqueName: \"kubernetes.io/projected/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-kube-api-access-8hdk8\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.490470 4813 scope.go:117] "RemoveContainer" containerID="90928cf06200bd034b239dd4beb66909b1ef8722ac8df18ff012dfba7270dec0" Mar 17 09:33:45 crc kubenswrapper[4813]: E0317 09:33:45.491152 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90928cf06200bd034b239dd4beb66909b1ef8722ac8df18ff012dfba7270dec0\": container with ID starting with 90928cf06200bd034b239dd4beb66909b1ef8722ac8df18ff012dfba7270dec0 not found: ID does not exist" containerID="90928cf06200bd034b239dd4beb66909b1ef8722ac8df18ff012dfba7270dec0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.491193 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90928cf06200bd034b239dd4beb66909b1ef8722ac8df18ff012dfba7270dec0"} err="failed to get container status \"90928cf06200bd034b239dd4beb66909b1ef8722ac8df18ff012dfba7270dec0\": rpc error: code = NotFound desc = could not find container \"90928cf06200bd034b239dd4beb66909b1ef8722ac8df18ff012dfba7270dec0\": container with ID starting with 90928cf06200bd034b239dd4beb66909b1ef8722ac8df18ff012dfba7270dec0 not found: ID does not exist" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.491220 4813 scope.go:117] "RemoveContainer" containerID="231fb631697d176a2411dfed49a192eac36fc630290640149cc808125041f064" Mar 17 09:33:45 crc kubenswrapper[4813]: E0317 09:33:45.491487 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"231fb631697d176a2411dfed49a192eac36fc630290640149cc808125041f064\": container with ID starting with 231fb631697d176a2411dfed49a192eac36fc630290640149cc808125041f064 not found: ID does not exist" containerID="231fb631697d176a2411dfed49a192eac36fc630290640149cc808125041f064" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.491571 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"231fb631697d176a2411dfed49a192eac36fc630290640149cc808125041f064"} err="failed to get container status \"231fb631697d176a2411dfed49a192eac36fc630290640149cc808125041f064\": rpc error: code = NotFound desc = could not find container \"231fb631697d176a2411dfed49a192eac36fc630290640149cc808125041f064\": container with ID starting with 231fb631697d176a2411dfed49a192eac36fc630290640149cc808125041f064 not found: ID does not exist" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.491674 4813 scope.go:117] "RemoveContainer" containerID="73117d1581b3b0162cf932507df289dd3c6ab947a79ae33d8c35fa36600444e4" Mar 17 09:33:45 crc kubenswrapper[4813]: E0317 09:33:45.492021 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73117d1581b3b0162cf932507df289dd3c6ab947a79ae33d8c35fa36600444e4\": container with ID starting with 73117d1581b3b0162cf932507df289dd3c6ab947a79ae33d8c35fa36600444e4 not found: ID does not exist" containerID="73117d1581b3b0162cf932507df289dd3c6ab947a79ae33d8c35fa36600444e4" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.492103 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73117d1581b3b0162cf932507df289dd3c6ab947a79ae33d8c35fa36600444e4"} err="failed to get container status \"73117d1581b3b0162cf932507df289dd3c6ab947a79ae33d8c35fa36600444e4\": rpc error: code = NotFound desc = could not find container \"73117d1581b3b0162cf932507df289dd3c6ab947a79ae33d8c35fa36600444e4\": container with ID starting with 73117d1581b3b0162cf932507df289dd3c6ab947a79ae33d8c35fa36600444e4 not found: ID does not exist" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.492243 4813 scope.go:117] "RemoveContainer" containerID="f6e0137118032ea186557ccee3fa1aaf40e2eef858ffe86dab662cb8c5111fd6" Mar 17 09:33:45 crc kubenswrapper[4813]: E0317 09:33:45.492780 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6e0137118032ea186557ccee3fa1aaf40e2eef858ffe86dab662cb8c5111fd6\": container with ID starting with f6e0137118032ea186557ccee3fa1aaf40e2eef858ffe86dab662cb8c5111fd6 not found: ID does not exist" containerID="f6e0137118032ea186557ccee3fa1aaf40e2eef858ffe86dab662cb8c5111fd6" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.492855 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6e0137118032ea186557ccee3fa1aaf40e2eef858ffe86dab662cb8c5111fd6"} err="failed to get container status \"f6e0137118032ea186557ccee3fa1aaf40e2eef858ffe86dab662cb8c5111fd6\": rpc error: code = NotFound desc = could not find container \"f6e0137118032ea186557ccee3fa1aaf40e2eef858ffe86dab662cb8c5111fd6\": container with ID starting with f6e0137118032ea186557ccee3fa1aaf40e2eef858ffe86dab662cb8c5111fd6 not found: ID does not exist" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.563207 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.563449 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.563650 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-run-httpd\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.563754 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-config-data\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.563896 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.564003 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-scripts\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.564226 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-log-httpd\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.564481 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hdk8\" (UniqueName: \"kubernetes.io/projected/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-kube-api-access-8hdk8\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.564276 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-run-httpd\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.564692 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-log-httpd\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.568541 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.570094 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.570539 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-scripts\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.572550 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.572668 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-config-data\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.582333 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hdk8\" (UniqueName: \"kubernetes.io/projected/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-kube-api-access-8hdk8\") pod \"ceilometer-0\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " pod="openstack/ceilometer-0" Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.773374 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:33:45 crc kubenswrapper[4813]: I0317 09:33:45.774073 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:33:46 crc kubenswrapper[4813]: I0317 09:33:46.288672 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:33:46 crc kubenswrapper[4813]: W0317 09:33:46.304112 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd17ec189_0bc4_40f9_a6b2_935f7ce4c2a4.slice/crio-4e5cd16e00641344b4cd5ca5a749ff7ebbf3468dc32744aa5e7e5b6b88dd61d0 WatchSource:0}: Error finding container 4e5cd16e00641344b4cd5ca5a749ff7ebbf3468dc32744aa5e7e5b6b88dd61d0: Status 404 returned error can't find the container with id 4e5cd16e00641344b4cd5ca5a749ff7ebbf3468dc32744aa5e7e5b6b88dd61d0 Mar 17 09:33:46 crc kubenswrapper[4813]: I0317 09:33:46.381322 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4","Type":"ContainerStarted","Data":"4e5cd16e00641344b4cd5ca5a749ff7ebbf3468dc32744aa5e7e5b6b88dd61d0"} Mar 17 09:33:46 crc kubenswrapper[4813]: I0317 09:33:46.744725 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98d79382-0d20-4868-b265-cbf83126f43d" path="/var/lib/kubelet/pods/98d79382-0d20-4868-b265-cbf83126f43d/volumes" Mar 17 09:33:47 crc kubenswrapper[4813]: I0317 09:33:47.400531 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4","Type":"ContainerStarted","Data":"bc8a22a0a5d1c2e6b6261596835dc225622463bdb8dc5b252449d641bea9b7a7"} Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:47.966036 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.140119 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-combined-ca-bundle\") pod \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\" (UID: \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\") " Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.141974 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-logs\") pod \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\" (UID: \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\") " Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.142442 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-logs" (OuterVolumeSpecName: "logs") pod "9e13c70a-9e05-4d99-a662-ac6cff2f3abc" (UID: "9e13c70a-9e05-4d99-a662-ac6cff2f3abc"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.142646 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qql8x\" (UniqueName: \"kubernetes.io/projected/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-kube-api-access-qql8x\") pod \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\" (UID: \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\") " Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.142722 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-config-data\") pod \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\" (UID: \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\") " Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.143900 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-logs\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.161636 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-kube-api-access-qql8x" (OuterVolumeSpecName: "kube-api-access-qql8x") pod "9e13c70a-9e05-4d99-a662-ac6cff2f3abc" (UID: "9e13c70a-9e05-4d99-a662-ac6cff2f3abc"). InnerVolumeSpecName "kube-api-access-qql8x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:33:48 crc kubenswrapper[4813]: E0317 09:33:48.177063 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-config-data podName:9e13c70a-9e05-4d99-a662-ac6cff2f3abc nodeName:}" failed. No retries permitted until 2026-03-17 09:33:48.67702942 +0000 UTC m=+1450.777832919 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config-data" (UniqueName: "kubernetes.io/secret/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-config-data") pod "9e13c70a-9e05-4d99-a662-ac6cff2f3abc" (UID: "9e13c70a-9e05-4d99-a662-ac6cff2f3abc") : error deleting /var/lib/kubelet/pods/9e13c70a-9e05-4d99-a662-ac6cff2f3abc/volume-subpaths: remove /var/lib/kubelet/pods/9e13c70a-9e05-4d99-a662-ac6cff2f3abc/volume-subpaths: no such file or directory Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.182781 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e13c70a-9e05-4d99-a662-ac6cff2f3abc" (UID: "9e13c70a-9e05-4d99-a662-ac6cff2f3abc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.245338 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.245380 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qql8x\" (UniqueName: \"kubernetes.io/projected/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-kube-api-access-qql8x\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.414181 4813 generic.go:334] "Generic (PLEG): container finished" podID="9e13c70a-9e05-4d99-a662-ac6cff2f3abc" containerID="ee37e13ddff7378536b800c5d4d189cffbc92a4388fc73492ed5f52ff2f7c3a4" exitCode=0 Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.414258 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.414228 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9e13c70a-9e05-4d99-a662-ac6cff2f3abc","Type":"ContainerDied","Data":"ee37e13ddff7378536b800c5d4d189cffbc92a4388fc73492ed5f52ff2f7c3a4"} Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.414471 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9e13c70a-9e05-4d99-a662-ac6cff2f3abc","Type":"ContainerDied","Data":"2c01de7864b85f0fddca34ef67a17da121788f3745742569446f04cefbbeaa08"} Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.414530 4813 scope.go:117] "RemoveContainer" containerID="ee37e13ddff7378536b800c5d4d189cffbc92a4388fc73492ed5f52ff2f7c3a4" Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.448314 4813 scope.go:117] "RemoveContainer" containerID="1b05ed8d644fe11cf3ceb7f433f9d73358cfe9ad0832c1b180007725a6d5d9ea" Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.475922 4813 scope.go:117] "RemoveContainer" containerID="ee37e13ddff7378536b800c5d4d189cffbc92a4388fc73492ed5f52ff2f7c3a4" Mar 17 09:33:48 crc kubenswrapper[4813]: E0317 09:33:48.476295 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ee37e13ddff7378536b800c5d4d189cffbc92a4388fc73492ed5f52ff2f7c3a4\": container with ID starting with ee37e13ddff7378536b800c5d4d189cffbc92a4388fc73492ed5f52ff2f7c3a4 not found: ID does not exist" containerID="ee37e13ddff7378536b800c5d4d189cffbc92a4388fc73492ed5f52ff2f7c3a4" Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.476325 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ee37e13ddff7378536b800c5d4d189cffbc92a4388fc73492ed5f52ff2f7c3a4"} err="failed to get container status \"ee37e13ddff7378536b800c5d4d189cffbc92a4388fc73492ed5f52ff2f7c3a4\": rpc error: code = NotFound desc = could not find container \"ee37e13ddff7378536b800c5d4d189cffbc92a4388fc73492ed5f52ff2f7c3a4\": container with ID starting with ee37e13ddff7378536b800c5d4d189cffbc92a4388fc73492ed5f52ff2f7c3a4 not found: ID does not exist" Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.476353 4813 scope.go:117] "RemoveContainer" containerID="1b05ed8d644fe11cf3ceb7f433f9d73358cfe9ad0832c1b180007725a6d5d9ea" Mar 17 09:33:48 crc kubenswrapper[4813]: E0317 09:33:48.476850 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1b05ed8d644fe11cf3ceb7f433f9d73358cfe9ad0832c1b180007725a6d5d9ea\": container with ID starting with 1b05ed8d644fe11cf3ceb7f433f9d73358cfe9ad0832c1b180007725a6d5d9ea not found: ID does not exist" containerID="1b05ed8d644fe11cf3ceb7f433f9d73358cfe9ad0832c1b180007725a6d5d9ea" Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.476874 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1b05ed8d644fe11cf3ceb7f433f9d73358cfe9ad0832c1b180007725a6d5d9ea"} err="failed to get container status \"1b05ed8d644fe11cf3ceb7f433f9d73358cfe9ad0832c1b180007725a6d5d9ea\": rpc error: code = NotFound desc = could not find container \"1b05ed8d644fe11cf3ceb7f433f9d73358cfe9ad0832c1b180007725a6d5d9ea\": container with ID starting with 1b05ed8d644fe11cf3ceb7f433f9d73358cfe9ad0832c1b180007725a6d5d9ea not found: ID does not exist" Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.755763 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-config-data\") pod \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\" (UID: \"9e13c70a-9e05-4d99-a662-ac6cff2f3abc\") " Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.774792 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-config-data" (OuterVolumeSpecName: "config-data") pod "9e13c70a-9e05-4d99-a662-ac6cff2f3abc" (UID: "9e13c70a-9e05-4d99-a662-ac6cff2f3abc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:48 crc kubenswrapper[4813]: I0317 09:33:48.858053 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e13c70a-9e05-4d99-a662-ac6cff2f3abc-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.138615 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.150435 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.161565 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 17 09:33:49 crc kubenswrapper[4813]: E0317 09:33:49.161991 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e13c70a-9e05-4d99-a662-ac6cff2f3abc" containerName="nova-api-api" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.162007 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e13c70a-9e05-4d99-a662-ac6cff2f3abc" containerName="nova-api-api" Mar 17 09:33:49 crc kubenswrapper[4813]: E0317 09:33:49.162025 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e13c70a-9e05-4d99-a662-ac6cff2f3abc" containerName="nova-api-log" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.162031 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e13c70a-9e05-4d99-a662-ac6cff2f3abc" containerName="nova-api-log" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.162197 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e13c70a-9e05-4d99-a662-ac6cff2f3abc" containerName="nova-api-api" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.162223 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e13c70a-9e05-4d99-a662-ac6cff2f3abc" containerName="nova-api-log" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.163143 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.171483 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.171543 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.171695 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.174176 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.265468 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-config-data\") pod \"nova-api-0\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " pod="openstack/nova-api-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.265763 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " pod="openstack/nova-api-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.265800 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2675d3f0-5304-46f5-a180-02c3e5801807-logs\") pod \"nova-api-0\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " pod="openstack/nova-api-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.265835 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qrcvx\" (UniqueName: \"kubernetes.io/projected/2675d3f0-5304-46f5-a180-02c3e5801807-kube-api-access-qrcvx\") pod \"nova-api-0\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " pod="openstack/nova-api-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.265884 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-public-tls-certs\") pod \"nova-api-0\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " pod="openstack/nova-api-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.265921 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " pod="openstack/nova-api-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.368252 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-config-data\") pod \"nova-api-0\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " pod="openstack/nova-api-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.368344 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " pod="openstack/nova-api-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.368365 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2675d3f0-5304-46f5-a180-02c3e5801807-logs\") pod \"nova-api-0\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " pod="openstack/nova-api-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.368418 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qrcvx\" (UniqueName: \"kubernetes.io/projected/2675d3f0-5304-46f5-a180-02c3e5801807-kube-api-access-qrcvx\") pod \"nova-api-0\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " pod="openstack/nova-api-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.368480 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-public-tls-certs\") pod \"nova-api-0\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " pod="openstack/nova-api-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.368536 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " pod="openstack/nova-api-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.368910 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2675d3f0-5304-46f5-a180-02c3e5801807-logs\") pod \"nova-api-0\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " pod="openstack/nova-api-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.373748 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-public-tls-certs\") pod \"nova-api-0\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " pod="openstack/nova-api-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.373959 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-config-data\") pod \"nova-api-0\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " pod="openstack/nova-api-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.375135 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " pod="openstack/nova-api-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.377346 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " pod="openstack/nova-api-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.386352 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qrcvx\" (UniqueName: \"kubernetes.io/projected/2675d3f0-5304-46f5-a180-02c3e5801807-kube-api-access-qrcvx\") pod \"nova-api-0\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " pod="openstack/nova-api-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.423716 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4","Type":"ContainerStarted","Data":"0853538319e6c3dcfbee6d0973f67eb14a7d98924b2e2f019dc6b1fee6f7e570"} Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.423761 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4","Type":"ContainerStarted","Data":"bf388c0ab3e2ca0cc012f999dace0ad43d8f0ead7b2dac1a63451cc83c5e7e1f"} Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.479711 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.703702 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.742130 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:49 crc kubenswrapper[4813]: I0317 09:33:49.986443 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 17 09:33:49 crc kubenswrapper[4813]: W0317 09:33:49.987971 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2675d3f0_5304_46f5_a180_02c3e5801807.slice/crio-91e1152980ab0a2663c9817dcd7328489fcd8340e6230790f367ce0fd1b66d23 WatchSource:0}: Error finding container 91e1152980ab0a2663c9817dcd7328489fcd8340e6230790f367ce0fd1b66d23: Status 404 returned error can't find the container with id 91e1152980ab0a2663c9817dcd7328489fcd8340e6230790f367ce0fd1b66d23 Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.433871 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2675d3f0-5304-46f5-a180-02c3e5801807","Type":"ContainerStarted","Data":"4bc22b1bdfb8d46891cec59520a76d55681e6037a82185f68ba1493eb27873cd"} Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.433927 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2675d3f0-5304-46f5-a180-02c3e5801807","Type":"ContainerStarted","Data":"529356a8e0a8ad13a119aae164983bba752d46726592a02e5ca846870ebadaa3"} Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.433940 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2675d3f0-5304-46f5-a180-02c3e5801807","Type":"ContainerStarted","Data":"91e1152980ab0a2663c9817dcd7328489fcd8340e6230790f367ce0fd1b66d23"} Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.449610 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.463485 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=1.463463633 podStartE2EDuration="1.463463633s" podCreationTimestamp="2026-03-17 09:33:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:33:50.451403594 +0000 UTC m=+1452.552207093" watchObservedRunningTime="2026-03-17 09:33:50.463463633 +0000 UTC m=+1452.564267132" Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.604408 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-v8vll"] Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.605539 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-v8vll" Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.611968 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.612159 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.619763 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-v8vll"] Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.692905 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g2zb\" (UniqueName: \"kubernetes.io/projected/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-kube-api-access-7g2zb\") pod \"nova-cell1-cell-mapping-v8vll\" (UID: \"6607c0ac-29d9-4cd9-9d95-9cfce3717c76\") " pod="openstack/nova-cell1-cell-mapping-v8vll" Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.692958 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-scripts\") pod \"nova-cell1-cell-mapping-v8vll\" (UID: \"6607c0ac-29d9-4cd9-9d95-9cfce3717c76\") " pod="openstack/nova-cell1-cell-mapping-v8vll" Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.692978 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-config-data\") pod \"nova-cell1-cell-mapping-v8vll\" (UID: \"6607c0ac-29d9-4cd9-9d95-9cfce3717c76\") " pod="openstack/nova-cell1-cell-mapping-v8vll" Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.693116 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-v8vll\" (UID: \"6607c0ac-29d9-4cd9-9d95-9cfce3717c76\") " pod="openstack/nova-cell1-cell-mapping-v8vll" Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.749223 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e13c70a-9e05-4d99-a662-ac6cff2f3abc" path="/var/lib/kubelet/pods/9e13c70a-9e05-4d99-a662-ac6cff2f3abc/volumes" Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.794844 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g2zb\" (UniqueName: \"kubernetes.io/projected/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-kube-api-access-7g2zb\") pod \"nova-cell1-cell-mapping-v8vll\" (UID: \"6607c0ac-29d9-4cd9-9d95-9cfce3717c76\") " pod="openstack/nova-cell1-cell-mapping-v8vll" Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.794921 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-scripts\") pod \"nova-cell1-cell-mapping-v8vll\" (UID: \"6607c0ac-29d9-4cd9-9d95-9cfce3717c76\") " pod="openstack/nova-cell1-cell-mapping-v8vll" Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.794942 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-config-data\") pod \"nova-cell1-cell-mapping-v8vll\" (UID: \"6607c0ac-29d9-4cd9-9d95-9cfce3717c76\") " pod="openstack/nova-cell1-cell-mapping-v8vll" Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.794967 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-v8vll\" (UID: \"6607c0ac-29d9-4cd9-9d95-9cfce3717c76\") " pod="openstack/nova-cell1-cell-mapping-v8vll" Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.799914 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-v8vll\" (UID: \"6607c0ac-29d9-4cd9-9d95-9cfce3717c76\") " pod="openstack/nova-cell1-cell-mapping-v8vll" Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.800188 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-config-data\") pod \"nova-cell1-cell-mapping-v8vll\" (UID: \"6607c0ac-29d9-4cd9-9d95-9cfce3717c76\") " pod="openstack/nova-cell1-cell-mapping-v8vll" Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.800966 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-scripts\") pod \"nova-cell1-cell-mapping-v8vll\" (UID: \"6607c0ac-29d9-4cd9-9d95-9cfce3717c76\") " pod="openstack/nova-cell1-cell-mapping-v8vll" Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.819941 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g2zb\" (UniqueName: \"kubernetes.io/projected/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-kube-api-access-7g2zb\") pod \"nova-cell1-cell-mapping-v8vll\" (UID: \"6607c0ac-29d9-4cd9-9d95-9cfce3717c76\") " pod="openstack/nova-cell1-cell-mapping-v8vll" Mar 17 09:33:50 crc kubenswrapper[4813]: I0317 09:33:50.931878 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-v8vll" Mar 17 09:33:50 crc kubenswrapper[4813]: E0317 09:33:50.987899 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod845fdf7b_a3b0_487b_9bcd_efcc3b2b8513.slice\": RecentStats: unable to find data in memory cache]" Mar 17 09:33:51 crc kubenswrapper[4813]: I0317 09:33:51.447201 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4","Type":"ContainerStarted","Data":"10443a6c2b3201425a0a7b6c0359aeede02db4f397cf67addda341995c6eb071"} Mar 17 09:33:51 crc kubenswrapper[4813]: I0317 09:33:51.447498 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" containerName="sg-core" containerID="cri-o://0853538319e6c3dcfbee6d0973f67eb14a7d98924b2e2f019dc6b1fee6f7e570" gracePeriod=30 Mar 17 09:33:51 crc kubenswrapper[4813]: I0317 09:33:51.447456 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" containerName="ceilometer-central-agent" containerID="cri-o://bc8a22a0a5d1c2e6b6261596835dc225622463bdb8dc5b252449d641bea9b7a7" gracePeriod=30 Mar 17 09:33:51 crc kubenswrapper[4813]: I0317 09:33:51.447573 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" containerName="ceilometer-notification-agent" containerID="cri-o://bf388c0ab3e2ca0cc012f999dace0ad43d8f0ead7b2dac1a63451cc83c5e7e1f" gracePeriod=30 Mar 17 09:33:51 crc kubenswrapper[4813]: I0317 09:33:51.447628 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" containerName="proxy-httpd" containerID="cri-o://10443a6c2b3201425a0a7b6c0359aeede02db4f397cf67addda341995c6eb071" gracePeriod=30 Mar 17 09:33:51 crc kubenswrapper[4813]: I0317 09:33:51.488446 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-v8vll"] Mar 17 09:33:51 crc kubenswrapper[4813]: W0317 09:33:51.490169 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6607c0ac_29d9_4cd9_9d95_9cfce3717c76.slice/crio-b27678e337a02ccbcfd1c32c46895db5317c0d06aae1979de450f3a91fb94db5 WatchSource:0}: Error finding container b27678e337a02ccbcfd1c32c46895db5317c0d06aae1979de450f3a91fb94db5: Status 404 returned error can't find the container with id b27678e337a02ccbcfd1c32c46895db5317c0d06aae1979de450f3a91fb94db5 Mar 17 09:33:51 crc kubenswrapper[4813]: I0317 09:33:51.497205 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.787872924 podStartE2EDuration="6.497182358s" podCreationTimestamp="2026-03-17 09:33:45 +0000 UTC" firstStartedPulling="2026-03-17 09:33:46.306880827 +0000 UTC m=+1448.407684326" lastFinishedPulling="2026-03-17 09:33:51.016190241 +0000 UTC m=+1453.116993760" observedRunningTime="2026-03-17 09:33:51.491729047 +0000 UTC m=+1453.592532586" watchObservedRunningTime="2026-03-17 09:33:51.497182358 +0000 UTC m=+1453.597985897" Mar 17 09:33:52 crc kubenswrapper[4813]: I0317 09:33:52.348586 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:33:52 crc kubenswrapper[4813]: I0317 09:33:52.466953 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-k9crm"] Mar 17 09:33:52 crc kubenswrapper[4813]: I0317 09:33:52.467386 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" podUID="7666a75c-63d7-4a2f-9280-4ec426a02358" containerName="dnsmasq-dns" containerID="cri-o://ae597b85c32ec4a1fd3b033d11a1426d69c4b8aa2f4964db75b4e7734c51fac1" gracePeriod=10 Mar 17 09:33:52 crc kubenswrapper[4813]: I0317 09:33:52.484222 4813 generic.go:334] "Generic (PLEG): container finished" podID="d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" containerID="10443a6c2b3201425a0a7b6c0359aeede02db4f397cf67addda341995c6eb071" exitCode=0 Mar 17 09:33:52 crc kubenswrapper[4813]: I0317 09:33:52.484262 4813 generic.go:334] "Generic (PLEG): container finished" podID="d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" containerID="0853538319e6c3dcfbee6d0973f67eb14a7d98924b2e2f019dc6b1fee6f7e570" exitCode=2 Mar 17 09:33:52 crc kubenswrapper[4813]: I0317 09:33:52.484273 4813 generic.go:334] "Generic (PLEG): container finished" podID="d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" containerID="bf388c0ab3e2ca0cc012f999dace0ad43d8f0ead7b2dac1a63451cc83c5e7e1f" exitCode=0 Mar 17 09:33:52 crc kubenswrapper[4813]: I0317 09:33:52.484322 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4","Type":"ContainerDied","Data":"10443a6c2b3201425a0a7b6c0359aeede02db4f397cf67addda341995c6eb071"} Mar 17 09:33:52 crc kubenswrapper[4813]: I0317 09:33:52.484349 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4","Type":"ContainerDied","Data":"0853538319e6c3dcfbee6d0973f67eb14a7d98924b2e2f019dc6b1fee6f7e570"} Mar 17 09:33:52 crc kubenswrapper[4813]: I0317 09:33:52.484360 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4","Type":"ContainerDied","Data":"bf388c0ab3e2ca0cc012f999dace0ad43d8f0ead7b2dac1a63451cc83c5e7e1f"} Mar 17 09:33:52 crc kubenswrapper[4813]: I0317 09:33:52.494979 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-v8vll" event={"ID":"6607c0ac-29d9-4cd9-9d95-9cfce3717c76","Type":"ContainerStarted","Data":"39837b8fda3965d754da6e2467aa1025f4b7f30d3903d9901da74fa873d3e1db"} Mar 17 09:33:52 crc kubenswrapper[4813]: I0317 09:33:52.495029 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-v8vll" event={"ID":"6607c0ac-29d9-4cd9-9d95-9cfce3717c76","Type":"ContainerStarted","Data":"b27678e337a02ccbcfd1c32c46895db5317c0d06aae1979de450f3a91fb94db5"} Mar 17 09:33:52 crc kubenswrapper[4813]: I0317 09:33:52.520766 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-v8vll" podStartSLOduration=2.520747267 podStartE2EDuration="2.520747267s" podCreationTimestamp="2026-03-17 09:33:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:33:52.507661896 +0000 UTC m=+1454.608465415" watchObservedRunningTime="2026-03-17 09:33:52.520747267 +0000 UTC m=+1454.621550766" Mar 17 09:33:52 crc kubenswrapper[4813]: I0317 09:33:52.981715 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.050520 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-ovsdbserver-nb\") pod \"7666a75c-63d7-4a2f-9280-4ec426a02358\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.050615 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-dns-swift-storage-0\") pod \"7666a75c-63d7-4a2f-9280-4ec426a02358\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.050665 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-ovsdbserver-sb\") pod \"7666a75c-63d7-4a2f-9280-4ec426a02358\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.050686 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-config\") pod \"7666a75c-63d7-4a2f-9280-4ec426a02358\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.050763 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-dns-svc\") pod \"7666a75c-63d7-4a2f-9280-4ec426a02358\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.050786 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v5xx6\" (UniqueName: \"kubernetes.io/projected/7666a75c-63d7-4a2f-9280-4ec426a02358-kube-api-access-v5xx6\") pod \"7666a75c-63d7-4a2f-9280-4ec426a02358\" (UID: \"7666a75c-63d7-4a2f-9280-4ec426a02358\") " Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.108482 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7666a75c-63d7-4a2f-9280-4ec426a02358-kube-api-access-v5xx6" (OuterVolumeSpecName: "kube-api-access-v5xx6") pod "7666a75c-63d7-4a2f-9280-4ec426a02358" (UID: "7666a75c-63d7-4a2f-9280-4ec426a02358"). InnerVolumeSpecName "kube-api-access-v5xx6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.167864 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v5xx6\" (UniqueName: \"kubernetes.io/projected/7666a75c-63d7-4a2f-9280-4ec426a02358-kube-api-access-v5xx6\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.204291 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7666a75c-63d7-4a2f-9280-4ec426a02358" (UID: "7666a75c-63d7-4a2f-9280-4ec426a02358"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.216583 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7666a75c-63d7-4a2f-9280-4ec426a02358" (UID: "7666a75c-63d7-4a2f-9280-4ec426a02358"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.224901 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7666a75c-63d7-4a2f-9280-4ec426a02358" (UID: "7666a75c-63d7-4a2f-9280-4ec426a02358"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.241159 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-config" (OuterVolumeSpecName: "config") pod "7666a75c-63d7-4a2f-9280-4ec426a02358" (UID: "7666a75c-63d7-4a2f-9280-4ec426a02358"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.253456 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7666a75c-63d7-4a2f-9280-4ec426a02358" (UID: "7666a75c-63d7-4a2f-9280-4ec426a02358"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.269818 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.269848 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.269864 4813 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.269873 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.269883 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7666a75c-63d7-4a2f-9280-4ec426a02358-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.506671 4813 generic.go:334] "Generic (PLEG): container finished" podID="7666a75c-63d7-4a2f-9280-4ec426a02358" containerID="ae597b85c32ec4a1fd3b033d11a1426d69c4b8aa2f4964db75b4e7734c51fac1" exitCode=0 Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.506748 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.506750 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" event={"ID":"7666a75c-63d7-4a2f-9280-4ec426a02358","Type":"ContainerDied","Data":"ae597b85c32ec4a1fd3b033d11a1426d69c4b8aa2f4964db75b4e7734c51fac1"} Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.507115 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-845d6d6f59-k9crm" event={"ID":"7666a75c-63d7-4a2f-9280-4ec426a02358","Type":"ContainerDied","Data":"59725a6479c946c1029cbff3f9dae15747e3bee03b822295a7d17081b71fda3d"} Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.507139 4813 scope.go:117] "RemoveContainer" containerID="ae597b85c32ec4a1fd3b033d11a1426d69c4b8aa2f4964db75b4e7734c51fac1" Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.538074 4813 scope.go:117] "RemoveContainer" containerID="d5b1a865da51567bf15798bce886e42ac5f2b60764cec2deb2ffced06ca72f62" Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.558170 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-k9crm"] Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.567496 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-845d6d6f59-k9crm"] Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.569028 4813 scope.go:117] "RemoveContainer" containerID="ae597b85c32ec4a1fd3b033d11a1426d69c4b8aa2f4964db75b4e7734c51fac1" Mar 17 09:33:53 crc kubenswrapper[4813]: E0317 09:33:53.569495 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae597b85c32ec4a1fd3b033d11a1426d69c4b8aa2f4964db75b4e7734c51fac1\": container with ID starting with ae597b85c32ec4a1fd3b033d11a1426d69c4b8aa2f4964db75b4e7734c51fac1 not found: ID does not exist" containerID="ae597b85c32ec4a1fd3b033d11a1426d69c4b8aa2f4964db75b4e7734c51fac1" Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.569531 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae597b85c32ec4a1fd3b033d11a1426d69c4b8aa2f4964db75b4e7734c51fac1"} err="failed to get container status \"ae597b85c32ec4a1fd3b033d11a1426d69c4b8aa2f4964db75b4e7734c51fac1\": rpc error: code = NotFound desc = could not find container \"ae597b85c32ec4a1fd3b033d11a1426d69c4b8aa2f4964db75b4e7734c51fac1\": container with ID starting with ae597b85c32ec4a1fd3b033d11a1426d69c4b8aa2f4964db75b4e7734c51fac1 not found: ID does not exist" Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.569581 4813 scope.go:117] "RemoveContainer" containerID="d5b1a865da51567bf15798bce886e42ac5f2b60764cec2deb2ffced06ca72f62" Mar 17 09:33:53 crc kubenswrapper[4813]: E0317 09:33:53.570098 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5b1a865da51567bf15798bce886e42ac5f2b60764cec2deb2ffced06ca72f62\": container with ID starting with d5b1a865da51567bf15798bce886e42ac5f2b60764cec2deb2ffced06ca72f62 not found: ID does not exist" containerID="d5b1a865da51567bf15798bce886e42ac5f2b60764cec2deb2ffced06ca72f62" Mar 17 09:33:53 crc kubenswrapper[4813]: I0317 09:33:53.570162 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5b1a865da51567bf15798bce886e42ac5f2b60764cec2deb2ffced06ca72f62"} err="failed to get container status \"d5b1a865da51567bf15798bce886e42ac5f2b60764cec2deb2ffced06ca72f62\": rpc error: code = NotFound desc = could not find container \"d5b1a865da51567bf15798bce886e42ac5f2b60764cec2deb2ffced06ca72f62\": container with ID starting with d5b1a865da51567bf15798bce886e42ac5f2b60764cec2deb2ffced06ca72f62 not found: ID does not exist" Mar 17 09:33:54 crc kubenswrapper[4813]: I0317 09:33:54.748637 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7666a75c-63d7-4a2f-9280-4ec426a02358" path="/var/lib/kubelet/pods/7666a75c-63d7-4a2f-9280-4ec426a02358/volumes" Mar 17 09:33:54 crc kubenswrapper[4813]: I0317 09:33:54.937831 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.103047 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8hdk8\" (UniqueName: \"kubernetes.io/projected/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-kube-api-access-8hdk8\") pod \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.103145 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-combined-ca-bundle\") pod \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.103311 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-run-httpd\") pod \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.103375 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-scripts\") pod \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.103417 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-sg-core-conf-yaml\") pod \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.103520 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-log-httpd\") pod \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.103579 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-config-data\") pod \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.103640 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-ceilometer-tls-certs\") pod \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\" (UID: \"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4\") " Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.105837 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" (UID: "d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.106338 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" (UID: "d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.113756 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-scripts" (OuterVolumeSpecName: "scripts") pod "d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" (UID: "d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.115806 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-kube-api-access-8hdk8" (OuterVolumeSpecName: "kube-api-access-8hdk8") pod "d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" (UID: "d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4"). InnerVolumeSpecName "kube-api-access-8hdk8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.142192 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" (UID: "d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.165285 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" (UID: "d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.206272 4813 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-run-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.206309 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.206327 4813 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.206346 4813 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-log-httpd\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.206363 4813 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.206381 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8hdk8\" (UniqueName: \"kubernetes.io/projected/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-kube-api-access-8hdk8\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.224834 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" (UID: "d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.267094 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-config-data" (OuterVolumeSpecName: "config-data") pod "d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" (UID: "d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.308797 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.308832 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.540889 4813 generic.go:334] "Generic (PLEG): container finished" podID="d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" containerID="bc8a22a0a5d1c2e6b6261596835dc225622463bdb8dc5b252449d641bea9b7a7" exitCode=0 Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.540998 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.541019 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4","Type":"ContainerDied","Data":"bc8a22a0a5d1c2e6b6261596835dc225622463bdb8dc5b252449d641bea9b7a7"} Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.541396 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4","Type":"ContainerDied","Data":"4e5cd16e00641344b4cd5ca5a749ff7ebbf3468dc32744aa5e7e5b6b88dd61d0"} Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.541427 4813 scope.go:117] "RemoveContainer" containerID="10443a6c2b3201425a0a7b6c0359aeede02db4f397cf67addda341995c6eb071" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.580825 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.593289 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.617079 4813 scope.go:117] "RemoveContainer" containerID="0853538319e6c3dcfbee6d0973f67eb14a7d98924b2e2f019dc6b1fee6f7e570" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.641281 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:33:55 crc kubenswrapper[4813]: E0317 09:33:55.641995 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" containerName="ceilometer-central-agent" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.642091 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" containerName="ceilometer-central-agent" Mar 17 09:33:55 crc kubenswrapper[4813]: E0317 09:33:55.642187 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" containerName="proxy-httpd" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.642262 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" containerName="proxy-httpd" Mar 17 09:33:55 crc kubenswrapper[4813]: E0317 09:33:55.642345 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7666a75c-63d7-4a2f-9280-4ec426a02358" containerName="init" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.642421 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7666a75c-63d7-4a2f-9280-4ec426a02358" containerName="init" Mar 17 09:33:55 crc kubenswrapper[4813]: E0317 09:33:55.642534 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" containerName="sg-core" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.642634 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" containerName="sg-core" Mar 17 09:33:55 crc kubenswrapper[4813]: E0317 09:33:55.642725 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7666a75c-63d7-4a2f-9280-4ec426a02358" containerName="dnsmasq-dns" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.642796 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="7666a75c-63d7-4a2f-9280-4ec426a02358" containerName="dnsmasq-dns" Mar 17 09:33:55 crc kubenswrapper[4813]: E0317 09:33:55.642880 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" containerName="ceilometer-notification-agent" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.642956 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" containerName="ceilometer-notification-agent" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.643420 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" containerName="ceilometer-notification-agent" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.643523 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="7666a75c-63d7-4a2f-9280-4ec426a02358" containerName="dnsmasq-dns" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.643613 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" containerName="ceilometer-central-agent" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.643697 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" containerName="proxy-httpd" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.643789 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" containerName="sg-core" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.646134 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.646423 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.648572 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.650793 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.650976 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.699267 4813 scope.go:117] "RemoveContainer" containerID="bf388c0ab3e2ca0cc012f999dace0ad43d8f0ead7b2dac1a63451cc83c5e7e1f" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.718314 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.718585 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.718778 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q6gcj\" (UniqueName: \"kubernetes.io/projected/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-kube-api-access-q6gcj\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.718925 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-run-httpd\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.719064 4813 scope.go:117] "RemoveContainer" containerID="bc8a22a0a5d1c2e6b6261596835dc225622463bdb8dc5b252449d641bea9b7a7" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.719085 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-config-data\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.719314 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.719362 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-log-httpd\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.719397 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-scripts\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.737110 4813 scope.go:117] "RemoveContainer" containerID="10443a6c2b3201425a0a7b6c0359aeede02db4f397cf67addda341995c6eb071" Mar 17 09:33:55 crc kubenswrapper[4813]: E0317 09:33:55.737719 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"10443a6c2b3201425a0a7b6c0359aeede02db4f397cf67addda341995c6eb071\": container with ID starting with 10443a6c2b3201425a0a7b6c0359aeede02db4f397cf67addda341995c6eb071 not found: ID does not exist" containerID="10443a6c2b3201425a0a7b6c0359aeede02db4f397cf67addda341995c6eb071" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.737773 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"10443a6c2b3201425a0a7b6c0359aeede02db4f397cf67addda341995c6eb071"} err="failed to get container status \"10443a6c2b3201425a0a7b6c0359aeede02db4f397cf67addda341995c6eb071\": rpc error: code = NotFound desc = could not find container \"10443a6c2b3201425a0a7b6c0359aeede02db4f397cf67addda341995c6eb071\": container with ID starting with 10443a6c2b3201425a0a7b6c0359aeede02db4f397cf67addda341995c6eb071 not found: ID does not exist" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.737801 4813 scope.go:117] "RemoveContainer" containerID="0853538319e6c3dcfbee6d0973f67eb14a7d98924b2e2f019dc6b1fee6f7e570" Mar 17 09:33:55 crc kubenswrapper[4813]: E0317 09:33:55.738156 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0853538319e6c3dcfbee6d0973f67eb14a7d98924b2e2f019dc6b1fee6f7e570\": container with ID starting with 0853538319e6c3dcfbee6d0973f67eb14a7d98924b2e2f019dc6b1fee6f7e570 not found: ID does not exist" containerID="0853538319e6c3dcfbee6d0973f67eb14a7d98924b2e2f019dc6b1fee6f7e570" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.738203 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0853538319e6c3dcfbee6d0973f67eb14a7d98924b2e2f019dc6b1fee6f7e570"} err="failed to get container status \"0853538319e6c3dcfbee6d0973f67eb14a7d98924b2e2f019dc6b1fee6f7e570\": rpc error: code = NotFound desc = could not find container \"0853538319e6c3dcfbee6d0973f67eb14a7d98924b2e2f019dc6b1fee6f7e570\": container with ID starting with 0853538319e6c3dcfbee6d0973f67eb14a7d98924b2e2f019dc6b1fee6f7e570 not found: ID does not exist" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.738221 4813 scope.go:117] "RemoveContainer" containerID="bf388c0ab3e2ca0cc012f999dace0ad43d8f0ead7b2dac1a63451cc83c5e7e1f" Mar 17 09:33:55 crc kubenswrapper[4813]: E0317 09:33:55.738582 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bf388c0ab3e2ca0cc012f999dace0ad43d8f0ead7b2dac1a63451cc83c5e7e1f\": container with ID starting with bf388c0ab3e2ca0cc012f999dace0ad43d8f0ead7b2dac1a63451cc83c5e7e1f not found: ID does not exist" containerID="bf388c0ab3e2ca0cc012f999dace0ad43d8f0ead7b2dac1a63451cc83c5e7e1f" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.738660 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bf388c0ab3e2ca0cc012f999dace0ad43d8f0ead7b2dac1a63451cc83c5e7e1f"} err="failed to get container status \"bf388c0ab3e2ca0cc012f999dace0ad43d8f0ead7b2dac1a63451cc83c5e7e1f\": rpc error: code = NotFound desc = could not find container \"bf388c0ab3e2ca0cc012f999dace0ad43d8f0ead7b2dac1a63451cc83c5e7e1f\": container with ID starting with bf388c0ab3e2ca0cc012f999dace0ad43d8f0ead7b2dac1a63451cc83c5e7e1f not found: ID does not exist" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.738700 4813 scope.go:117] "RemoveContainer" containerID="bc8a22a0a5d1c2e6b6261596835dc225622463bdb8dc5b252449d641bea9b7a7" Mar 17 09:33:55 crc kubenswrapper[4813]: E0317 09:33:55.739117 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc8a22a0a5d1c2e6b6261596835dc225622463bdb8dc5b252449d641bea9b7a7\": container with ID starting with bc8a22a0a5d1c2e6b6261596835dc225622463bdb8dc5b252449d641bea9b7a7 not found: ID does not exist" containerID="bc8a22a0a5d1c2e6b6261596835dc225622463bdb8dc5b252449d641bea9b7a7" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.739197 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc8a22a0a5d1c2e6b6261596835dc225622463bdb8dc5b252449d641bea9b7a7"} err="failed to get container status \"bc8a22a0a5d1c2e6b6261596835dc225622463bdb8dc5b252449d641bea9b7a7\": rpc error: code = NotFound desc = could not find container \"bc8a22a0a5d1c2e6b6261596835dc225622463bdb8dc5b252449d641bea9b7a7\": container with ID starting with bc8a22a0a5d1c2e6b6261596835dc225622463bdb8dc5b252449d641bea9b7a7 not found: ID does not exist" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.821735 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q6gcj\" (UniqueName: \"kubernetes.io/projected/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-kube-api-access-q6gcj\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.821841 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-run-httpd\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.821927 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-config-data\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.822000 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.822039 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-log-httpd\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.822071 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-scripts\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.822173 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.822272 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.823505 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-run-httpd\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.825235 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-log-httpd\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.828578 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.829380 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-scripts\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.829626 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-config-data\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.831320 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.842952 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q6gcj\" (UniqueName: \"kubernetes.io/projected/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-kube-api-access-q6gcj\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.851252 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea9cbc63-ff93-4342-ab4c-7ddf7bf78577-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577\") " pod="openstack/ceilometer-0" Mar 17 09:33:55 crc kubenswrapper[4813]: I0317 09:33:55.998890 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Mar 17 09:33:56 crc kubenswrapper[4813]: I0317 09:33:56.460590 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Mar 17 09:33:56 crc kubenswrapper[4813]: W0317 09:33:56.462390 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea9cbc63_ff93_4342_ab4c_7ddf7bf78577.slice/crio-098050753ac87a5682f124de5797f4277a202316494de1b7cdb5625c467fb2de WatchSource:0}: Error finding container 098050753ac87a5682f124de5797f4277a202316494de1b7cdb5625c467fb2de: Status 404 returned error can't find the container with id 098050753ac87a5682f124de5797f4277a202316494de1b7cdb5625c467fb2de Mar 17 09:33:56 crc kubenswrapper[4813]: I0317 09:33:56.555033 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577","Type":"ContainerStarted","Data":"098050753ac87a5682f124de5797f4277a202316494de1b7cdb5625c467fb2de"} Mar 17 09:33:56 crc kubenswrapper[4813]: I0317 09:33:56.557654 4813 generic.go:334] "Generic (PLEG): container finished" podID="6607c0ac-29d9-4cd9-9d95-9cfce3717c76" containerID="39837b8fda3965d754da6e2467aa1025f4b7f30d3903d9901da74fa873d3e1db" exitCode=0 Mar 17 09:33:56 crc kubenswrapper[4813]: I0317 09:33:56.557749 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-v8vll" event={"ID":"6607c0ac-29d9-4cd9-9d95-9cfce3717c76","Type":"ContainerDied","Data":"39837b8fda3965d754da6e2467aa1025f4b7f30d3903d9901da74fa873d3e1db"} Mar 17 09:33:56 crc kubenswrapper[4813]: I0317 09:33:56.741684 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4" path="/var/lib/kubelet/pods/d17ec189-0bc4-40f9-a6b2-935f7ce4c2a4/volumes" Mar 17 09:33:57 crc kubenswrapper[4813]: I0317 09:33:57.584765 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577","Type":"ContainerStarted","Data":"cf25d6dc68a0cf235f0ed26bcd3eec65b77a593f1e614e149b1b40b62f1d1b82"} Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.165530 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-v8vll" Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.266307 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-scripts\") pod \"6607c0ac-29d9-4cd9-9d95-9cfce3717c76\" (UID: \"6607c0ac-29d9-4cd9-9d95-9cfce3717c76\") " Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.266390 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-combined-ca-bundle\") pod \"6607c0ac-29d9-4cd9-9d95-9cfce3717c76\" (UID: \"6607c0ac-29d9-4cd9-9d95-9cfce3717c76\") " Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.266445 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7g2zb\" (UniqueName: \"kubernetes.io/projected/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-kube-api-access-7g2zb\") pod \"6607c0ac-29d9-4cd9-9d95-9cfce3717c76\" (UID: \"6607c0ac-29d9-4cd9-9d95-9cfce3717c76\") " Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.266463 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-config-data\") pod \"6607c0ac-29d9-4cd9-9d95-9cfce3717c76\" (UID: \"6607c0ac-29d9-4cd9-9d95-9cfce3717c76\") " Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.272738 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-kube-api-access-7g2zb" (OuterVolumeSpecName: "kube-api-access-7g2zb") pod "6607c0ac-29d9-4cd9-9d95-9cfce3717c76" (UID: "6607c0ac-29d9-4cd9-9d95-9cfce3717c76"). InnerVolumeSpecName "kube-api-access-7g2zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.273106 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-scripts" (OuterVolumeSpecName: "scripts") pod "6607c0ac-29d9-4cd9-9d95-9cfce3717c76" (UID: "6607c0ac-29d9-4cd9-9d95-9cfce3717c76"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.310564 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-config-data" (OuterVolumeSpecName: "config-data") pod "6607c0ac-29d9-4cd9-9d95-9cfce3717c76" (UID: "6607c0ac-29d9-4cd9-9d95-9cfce3717c76"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.330699 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6607c0ac-29d9-4cd9-9d95-9cfce3717c76" (UID: "6607c0ac-29d9-4cd9-9d95-9cfce3717c76"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.368194 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.368249 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7g2zb\" (UniqueName: \"kubernetes.io/projected/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-kube-api-access-7g2zb\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.368261 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.368269 4813 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6607c0ac-29d9-4cd9-9d95-9cfce3717c76-scripts\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.602493 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577","Type":"ContainerStarted","Data":"bd771aab250a6aee24ec985045ad57fb8adaa024c9a9afa895e9e87c4f71aae3"} Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.605219 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-v8vll" event={"ID":"6607c0ac-29d9-4cd9-9d95-9cfce3717c76","Type":"ContainerDied","Data":"b27678e337a02ccbcfd1c32c46895db5317c0d06aae1979de450f3a91fb94db5"} Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.605256 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b27678e337a02ccbcfd1c32c46895db5317c0d06aae1979de450f3a91fb94db5" Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.605316 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-v8vll" Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.771525 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.771932 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2675d3f0-5304-46f5-a180-02c3e5801807" containerName="nova-api-log" containerID="cri-o://529356a8e0a8ad13a119aae164983bba752d46726592a02e5ca846870ebadaa3" gracePeriod=30 Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.772101 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2675d3f0-5304-46f5-a180-02c3e5801807" containerName="nova-api-api" containerID="cri-o://4bc22b1bdfb8d46891cec59520a76d55681e6037a82185f68ba1493eb27873cd" gracePeriod=30 Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.816652 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.816875 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="69e2e92a-5161-4a09-b59d-464cb4490ff0" containerName="nova-scheduler-scheduler" containerID="cri-o://0b9a534ff3bf6cd920494816e2d55e49a081e9a72e266ef15de661be98e603b2" gracePeriod=30 Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.832582 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.833093 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4b69ffd2-16ba-45af-b3c9-d4f8b3effebe" containerName="nova-metadata-log" containerID="cri-o://a44bfd63612dfab25234155cad9e67aa869107fcf44e75fc72bab3b9bef5eb28" gracePeriod=30 Mar 17 09:33:58 crc kubenswrapper[4813]: I0317 09:33:58.833574 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="4b69ffd2-16ba-45af-b3c9-d4f8b3effebe" containerName="nova-metadata-metadata" containerID="cri-o://a2de365165a9e93f392da0a42a918ce0e6cf111a582c5b174f55c44c7d966791" gracePeriod=30 Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.541686 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.627671 4813 generic.go:334] "Generic (PLEG): container finished" podID="2675d3f0-5304-46f5-a180-02c3e5801807" containerID="4bc22b1bdfb8d46891cec59520a76d55681e6037a82185f68ba1493eb27873cd" exitCode=0 Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.627713 4813 generic.go:334] "Generic (PLEG): container finished" podID="2675d3f0-5304-46f5-a180-02c3e5801807" containerID="529356a8e0a8ad13a119aae164983bba752d46726592a02e5ca846870ebadaa3" exitCode=143 Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.627727 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.627778 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2675d3f0-5304-46f5-a180-02c3e5801807","Type":"ContainerDied","Data":"4bc22b1bdfb8d46891cec59520a76d55681e6037a82185f68ba1493eb27873cd"} Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.627808 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2675d3f0-5304-46f5-a180-02c3e5801807","Type":"ContainerDied","Data":"529356a8e0a8ad13a119aae164983bba752d46726592a02e5ca846870ebadaa3"} Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.627823 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2675d3f0-5304-46f5-a180-02c3e5801807","Type":"ContainerDied","Data":"91e1152980ab0a2663c9817dcd7328489fcd8340e6230790f367ce0fd1b66d23"} Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.627839 4813 scope.go:117] "RemoveContainer" containerID="4bc22b1bdfb8d46891cec59520a76d55681e6037a82185f68ba1493eb27873cd" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.632516 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577","Type":"ContainerStarted","Data":"9fa4a33071231b9ff294f8301993ad05ec7bcaffd316a054b3c734cb52b46b8b"} Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.635565 4813 generic.go:334] "Generic (PLEG): container finished" podID="4b69ffd2-16ba-45af-b3c9-d4f8b3effebe" containerID="a44bfd63612dfab25234155cad9e67aa869107fcf44e75fc72bab3b9bef5eb28" exitCode=143 Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.635611 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe","Type":"ContainerDied","Data":"a44bfd63612dfab25234155cad9e67aa869107fcf44e75fc72bab3b9bef5eb28"} Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.648580 4813 scope.go:117] "RemoveContainer" containerID="529356a8e0a8ad13a119aae164983bba752d46726592a02e5ca846870ebadaa3" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.666364 4813 scope.go:117] "RemoveContainer" containerID="4bc22b1bdfb8d46891cec59520a76d55681e6037a82185f68ba1493eb27873cd" Mar 17 09:33:59 crc kubenswrapper[4813]: E0317 09:33:59.666813 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bc22b1bdfb8d46891cec59520a76d55681e6037a82185f68ba1493eb27873cd\": container with ID starting with 4bc22b1bdfb8d46891cec59520a76d55681e6037a82185f68ba1493eb27873cd not found: ID does not exist" containerID="4bc22b1bdfb8d46891cec59520a76d55681e6037a82185f68ba1493eb27873cd" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.666860 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bc22b1bdfb8d46891cec59520a76d55681e6037a82185f68ba1493eb27873cd"} err="failed to get container status \"4bc22b1bdfb8d46891cec59520a76d55681e6037a82185f68ba1493eb27873cd\": rpc error: code = NotFound desc = could not find container \"4bc22b1bdfb8d46891cec59520a76d55681e6037a82185f68ba1493eb27873cd\": container with ID starting with 4bc22b1bdfb8d46891cec59520a76d55681e6037a82185f68ba1493eb27873cd not found: ID does not exist" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.666884 4813 scope.go:117] "RemoveContainer" containerID="529356a8e0a8ad13a119aae164983bba752d46726592a02e5ca846870ebadaa3" Mar 17 09:33:59 crc kubenswrapper[4813]: E0317 09:33:59.667178 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"529356a8e0a8ad13a119aae164983bba752d46726592a02e5ca846870ebadaa3\": container with ID starting with 529356a8e0a8ad13a119aae164983bba752d46726592a02e5ca846870ebadaa3 not found: ID does not exist" containerID="529356a8e0a8ad13a119aae164983bba752d46726592a02e5ca846870ebadaa3" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.667199 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"529356a8e0a8ad13a119aae164983bba752d46726592a02e5ca846870ebadaa3"} err="failed to get container status \"529356a8e0a8ad13a119aae164983bba752d46726592a02e5ca846870ebadaa3\": rpc error: code = NotFound desc = could not find container \"529356a8e0a8ad13a119aae164983bba752d46726592a02e5ca846870ebadaa3\": container with ID starting with 529356a8e0a8ad13a119aae164983bba752d46726592a02e5ca846870ebadaa3 not found: ID does not exist" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.667212 4813 scope.go:117] "RemoveContainer" containerID="4bc22b1bdfb8d46891cec59520a76d55681e6037a82185f68ba1493eb27873cd" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.667452 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bc22b1bdfb8d46891cec59520a76d55681e6037a82185f68ba1493eb27873cd"} err="failed to get container status \"4bc22b1bdfb8d46891cec59520a76d55681e6037a82185f68ba1493eb27873cd\": rpc error: code = NotFound desc = could not find container \"4bc22b1bdfb8d46891cec59520a76d55681e6037a82185f68ba1493eb27873cd\": container with ID starting with 4bc22b1bdfb8d46891cec59520a76d55681e6037a82185f68ba1493eb27873cd not found: ID does not exist" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.667471 4813 scope.go:117] "RemoveContainer" containerID="529356a8e0a8ad13a119aae164983bba752d46726592a02e5ca846870ebadaa3" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.667830 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"529356a8e0a8ad13a119aae164983bba752d46726592a02e5ca846870ebadaa3"} err="failed to get container status \"529356a8e0a8ad13a119aae164983bba752d46726592a02e5ca846870ebadaa3\": rpc error: code = NotFound desc = could not find container \"529356a8e0a8ad13a119aae164983bba752d46726592a02e5ca846870ebadaa3\": container with ID starting with 529356a8e0a8ad13a119aae164983bba752d46726592a02e5ca846870ebadaa3 not found: ID does not exist" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.689810 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qrcvx\" (UniqueName: \"kubernetes.io/projected/2675d3f0-5304-46f5-a180-02c3e5801807-kube-api-access-qrcvx\") pod \"2675d3f0-5304-46f5-a180-02c3e5801807\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.689905 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-combined-ca-bundle\") pod \"2675d3f0-5304-46f5-a180-02c3e5801807\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.690009 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-public-tls-certs\") pod \"2675d3f0-5304-46f5-a180-02c3e5801807\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.690062 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2675d3f0-5304-46f5-a180-02c3e5801807-logs\") pod \"2675d3f0-5304-46f5-a180-02c3e5801807\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.690110 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-internal-tls-certs\") pod \"2675d3f0-5304-46f5-a180-02c3e5801807\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.690137 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-config-data\") pod \"2675d3f0-5304-46f5-a180-02c3e5801807\" (UID: \"2675d3f0-5304-46f5-a180-02c3e5801807\") " Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.691349 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2675d3f0-5304-46f5-a180-02c3e5801807-logs" (OuterVolumeSpecName: "logs") pod "2675d3f0-5304-46f5-a180-02c3e5801807" (UID: "2675d3f0-5304-46f5-a180-02c3e5801807"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.695809 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2675d3f0-5304-46f5-a180-02c3e5801807-kube-api-access-qrcvx" (OuterVolumeSpecName: "kube-api-access-qrcvx") pod "2675d3f0-5304-46f5-a180-02c3e5801807" (UID: "2675d3f0-5304-46f5-a180-02c3e5801807"). InnerVolumeSpecName "kube-api-access-qrcvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.716767 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2675d3f0-5304-46f5-a180-02c3e5801807" (UID: "2675d3f0-5304-46f5-a180-02c3e5801807"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.719905 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-config-data" (OuterVolumeSpecName: "config-data") pod "2675d3f0-5304-46f5-a180-02c3e5801807" (UID: "2675d3f0-5304-46f5-a180-02c3e5801807"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:59 crc kubenswrapper[4813]: E0317 09:33:59.735956 4813 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0b9a534ff3bf6cd920494816e2d55e49a081e9a72e266ef15de661be98e603b2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 17 09:33:59 crc kubenswrapper[4813]: E0317 09:33:59.738102 4813 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0b9a534ff3bf6cd920494816e2d55e49a081e9a72e266ef15de661be98e603b2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.740896 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "2675d3f0-5304-46f5-a180-02c3e5801807" (UID: "2675d3f0-5304-46f5-a180-02c3e5801807"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:59 crc kubenswrapper[4813]: E0317 09:33:59.740961 4813 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="0b9a534ff3bf6cd920494816e2d55e49a081e9a72e266ef15de661be98e603b2" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Mar 17 09:33:59 crc kubenswrapper[4813]: E0317 09:33:59.741022 4813 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="69e2e92a-5161-4a09-b59d-464cb4490ff0" containerName="nova-scheduler-scheduler" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.748391 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "2675d3f0-5304-46f5-a180-02c3e5801807" (UID: "2675d3f0-5304-46f5-a180-02c3e5801807"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.791733 4813 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.792459 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.792502 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qrcvx\" (UniqueName: \"kubernetes.io/projected/2675d3f0-5304-46f5-a180-02c3e5801807-kube-api-access-qrcvx\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.792518 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.792531 4813 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2675d3f0-5304-46f5-a180-02c3e5801807-public-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.792544 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2675d3f0-5304-46f5-a180-02c3e5801807-logs\") on node \"crc\" DevicePath \"\"" Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.969374 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Mar 17 09:33:59 crc kubenswrapper[4813]: I0317 09:33:59.981500 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.003469 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Mar 17 09:34:00 crc kubenswrapper[4813]: E0317 09:34:00.004249 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2675d3f0-5304-46f5-a180-02c3e5801807" containerName="nova-api-api" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.004273 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="2675d3f0-5304-46f5-a180-02c3e5801807" containerName="nova-api-api" Mar 17 09:34:00 crc kubenswrapper[4813]: E0317 09:34:00.004305 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2675d3f0-5304-46f5-a180-02c3e5801807" containerName="nova-api-log" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.004314 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="2675d3f0-5304-46f5-a180-02c3e5801807" containerName="nova-api-log" Mar 17 09:34:00 crc kubenswrapper[4813]: E0317 09:34:00.004335 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6607c0ac-29d9-4cd9-9d95-9cfce3717c76" containerName="nova-manage" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.004344 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6607c0ac-29d9-4cd9-9d95-9cfce3717c76" containerName="nova-manage" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.004570 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="6607c0ac-29d9-4cd9-9d95-9cfce3717c76" containerName="nova-manage" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.004629 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="2675d3f0-5304-46f5-a180-02c3e5801807" containerName="nova-api-api" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.004649 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="2675d3f0-5304-46f5-a180-02c3e5801807" containerName="nova-api-log" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.005887 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.008226 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.008531 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.008694 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.014138 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.142224 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562334-bs752"] Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.143682 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562334-bs752" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.145369 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.147155 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.147359 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.151844 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562334-bs752"] Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.198584 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04ea2c77-25c5-45fc-a375-cf04b185381c-logs\") pod \"nova-api-0\" (UID: \"04ea2c77-25c5-45fc-a375-cf04b185381c\") " pod="openstack/nova-api-0" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.198708 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04ea2c77-25c5-45fc-a375-cf04b185381c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"04ea2c77-25c5-45fc-a375-cf04b185381c\") " pod="openstack/nova-api-0" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.198775 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04ea2c77-25c5-45fc-a375-cf04b185381c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"04ea2c77-25c5-45fc-a375-cf04b185381c\") " pod="openstack/nova-api-0" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.198796 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/04ea2c77-25c5-45fc-a375-cf04b185381c-public-tls-certs\") pod \"nova-api-0\" (UID: \"04ea2c77-25c5-45fc-a375-cf04b185381c\") " pod="openstack/nova-api-0" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.198826 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxzcb\" (UniqueName: \"kubernetes.io/projected/04ea2c77-25c5-45fc-a375-cf04b185381c-kube-api-access-qxzcb\") pod \"nova-api-0\" (UID: \"04ea2c77-25c5-45fc-a375-cf04b185381c\") " pod="openstack/nova-api-0" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.198894 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04ea2c77-25c5-45fc-a375-cf04b185381c-config-data\") pod \"nova-api-0\" (UID: \"04ea2c77-25c5-45fc-a375-cf04b185381c\") " pod="openstack/nova-api-0" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.300496 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04ea2c77-25c5-45fc-a375-cf04b185381c-logs\") pod \"nova-api-0\" (UID: \"04ea2c77-25c5-45fc-a375-cf04b185381c\") " pod="openstack/nova-api-0" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.300561 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04ea2c77-25c5-45fc-a375-cf04b185381c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"04ea2c77-25c5-45fc-a375-cf04b185381c\") " pod="openstack/nova-api-0" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.300647 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxz47\" (UniqueName: \"kubernetes.io/projected/d363247c-cdbf-49cc-b332-e87f849d9fee-kube-api-access-jxz47\") pod \"auto-csr-approver-29562334-bs752\" (UID: \"d363247c-cdbf-49cc-b332-e87f849d9fee\") " pod="openshift-infra/auto-csr-approver-29562334-bs752" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.300694 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04ea2c77-25c5-45fc-a375-cf04b185381c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"04ea2c77-25c5-45fc-a375-cf04b185381c\") " pod="openstack/nova-api-0" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.300718 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/04ea2c77-25c5-45fc-a375-cf04b185381c-public-tls-certs\") pod \"nova-api-0\" (UID: \"04ea2c77-25c5-45fc-a375-cf04b185381c\") " pod="openstack/nova-api-0" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.300753 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxzcb\" (UniqueName: \"kubernetes.io/projected/04ea2c77-25c5-45fc-a375-cf04b185381c-kube-api-access-qxzcb\") pod \"nova-api-0\" (UID: \"04ea2c77-25c5-45fc-a375-cf04b185381c\") " pod="openstack/nova-api-0" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.300829 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04ea2c77-25c5-45fc-a375-cf04b185381c-config-data\") pod \"nova-api-0\" (UID: \"04ea2c77-25c5-45fc-a375-cf04b185381c\") " pod="openstack/nova-api-0" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.301064 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/04ea2c77-25c5-45fc-a375-cf04b185381c-logs\") pod \"nova-api-0\" (UID: \"04ea2c77-25c5-45fc-a375-cf04b185381c\") " pod="openstack/nova-api-0" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.304848 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/04ea2c77-25c5-45fc-a375-cf04b185381c-config-data\") pod \"nova-api-0\" (UID: \"04ea2c77-25c5-45fc-a375-cf04b185381c\") " pod="openstack/nova-api-0" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.307932 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/04ea2c77-25c5-45fc-a375-cf04b185381c-public-tls-certs\") pod \"nova-api-0\" (UID: \"04ea2c77-25c5-45fc-a375-cf04b185381c\") " pod="openstack/nova-api-0" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.313930 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/04ea2c77-25c5-45fc-a375-cf04b185381c-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"04ea2c77-25c5-45fc-a375-cf04b185381c\") " pod="openstack/nova-api-0" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.319931 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/04ea2c77-25c5-45fc-a375-cf04b185381c-internal-tls-certs\") pod \"nova-api-0\" (UID: \"04ea2c77-25c5-45fc-a375-cf04b185381c\") " pod="openstack/nova-api-0" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.322272 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxzcb\" (UniqueName: \"kubernetes.io/projected/04ea2c77-25c5-45fc-a375-cf04b185381c-kube-api-access-qxzcb\") pod \"nova-api-0\" (UID: \"04ea2c77-25c5-45fc-a375-cf04b185381c\") " pod="openstack/nova-api-0" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.332831 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.402404 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxz47\" (UniqueName: \"kubernetes.io/projected/d363247c-cdbf-49cc-b332-e87f849d9fee-kube-api-access-jxz47\") pod \"auto-csr-approver-29562334-bs752\" (UID: \"d363247c-cdbf-49cc-b332-e87f849d9fee\") " pod="openshift-infra/auto-csr-approver-29562334-bs752" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.422469 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxz47\" (UniqueName: \"kubernetes.io/projected/d363247c-cdbf-49cc-b332-e87f849d9fee-kube-api-access-jxz47\") pod \"auto-csr-approver-29562334-bs752\" (UID: \"d363247c-cdbf-49cc-b332-e87f849d9fee\") " pod="openshift-infra/auto-csr-approver-29562334-bs752" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.495638 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562334-bs752" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.656143 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"ea9cbc63-ff93-4342-ab4c-7ddf7bf78577","Type":"ContainerStarted","Data":"054f111125add619c82ef8ba60a40f16799cb913b6fc5200e2f9181969cf65d2"} Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.657375 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.675663 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.790876535 podStartE2EDuration="5.675644594s" podCreationTimestamp="2026-03-17 09:33:55 +0000 UTC" firstStartedPulling="2026-03-17 09:33:56.466044495 +0000 UTC m=+1458.566848034" lastFinishedPulling="2026-03-17 09:34:00.350812594 +0000 UTC m=+1462.451616093" observedRunningTime="2026-03-17 09:34:00.675424427 +0000 UTC m=+1462.776227926" watchObservedRunningTime="2026-03-17 09:34:00.675644594 +0000 UTC m=+1462.776448093" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.740491 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2675d3f0-5304-46f5-a180-02c3e5801807" path="/var/lib/kubelet/pods/2675d3f0-5304-46f5-a180-02c3e5801807/volumes" Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.835276 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Mar 17 09:34:00 crc kubenswrapper[4813]: I0317 09:34:00.996470 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562334-bs752"] Mar 17 09:34:00 crc kubenswrapper[4813]: W0317 09:34:00.999025 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd363247c_cdbf_49cc_b332_e87f849d9fee.slice/crio-01d0511642c2f88c9f7f4e431d1a5c7e391c6ee7bfa8f4fb45a565ef41423641 WatchSource:0}: Error finding container 01d0511642c2f88c9f7f4e431d1a5c7e391c6ee7bfa8f4fb45a565ef41423641: Status 404 returned error can't find the container with id 01d0511642c2f88c9f7f4e431d1a5c7e391c6ee7bfa8f4fb45a565ef41423641 Mar 17 09:34:01 crc kubenswrapper[4813]: E0317 09:34:01.258062 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod845fdf7b_a3b0_487b_9bcd_efcc3b2b8513.slice\": RecentStats: unable to find data in memory cache]" Mar 17 09:34:01 crc kubenswrapper[4813]: I0317 09:34:01.664113 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"04ea2c77-25c5-45fc-a375-cf04b185381c","Type":"ContainerStarted","Data":"3eb55e176ea2b1145de2ac4ca331f7356c2b4e86232651d960c1e5fd997d4069"} Mar 17 09:34:01 crc kubenswrapper[4813]: I0317 09:34:01.664158 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"04ea2c77-25c5-45fc-a375-cf04b185381c","Type":"ContainerStarted","Data":"62351a40acbf5e115fea670df7e8df816f7b5fbccd5d561d02939b5bf79bf09b"} Mar 17 09:34:01 crc kubenswrapper[4813]: I0317 09:34:01.664168 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"04ea2c77-25c5-45fc-a375-cf04b185381c","Type":"ContainerStarted","Data":"0ea2a51c09d026e45f98e461e0978eca5a9ad5e6d64eb8275967c602e6eef446"} Mar 17 09:34:01 crc kubenswrapper[4813]: I0317 09:34:01.665587 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562334-bs752" event={"ID":"d363247c-cdbf-49cc-b332-e87f849d9fee","Type":"ContainerStarted","Data":"01d0511642c2f88c9f7f4e431d1a5c7e391c6ee7bfa8f4fb45a565ef41423641"} Mar 17 09:34:01 crc kubenswrapper[4813]: I0317 09:34:01.683890 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.68386306 podStartE2EDuration="2.68386306s" podCreationTimestamp="2026-03-17 09:33:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:34:01.681210807 +0000 UTC m=+1463.782014306" watchObservedRunningTime="2026-03-17 09:34:01.68386306 +0000 UTC m=+1463.784666569" Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.676758 4813 generic.go:334] "Generic (PLEG): container finished" podID="d363247c-cdbf-49cc-b332-e87f849d9fee" containerID="e4a14c386c4f92a231421b0f9a45151032bd59a580677087d5c334994723de64" exitCode=0 Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.676915 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562334-bs752" event={"ID":"d363247c-cdbf-49cc-b332-e87f849d9fee","Type":"ContainerDied","Data":"e4a14c386c4f92a231421b0f9a45151032bd59a580677087d5c334994723de64"} Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.680550 4813 generic.go:334] "Generic (PLEG): container finished" podID="4b69ffd2-16ba-45af-b3c9-d4f8b3effebe" containerID="a2de365165a9e93f392da0a42a918ce0e6cf111a582c5b174f55c44c7d966791" exitCode=0 Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.681633 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe","Type":"ContainerDied","Data":"a2de365165a9e93f392da0a42a918ce0e6cf111a582c5b174f55c44c7d966791"} Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.681683 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe","Type":"ContainerDied","Data":"928c1595632eaec6df4ebb7df9f1120bf3779a1a79a9d5774416f7f78fc34048"} Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.681703 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="928c1595632eaec6df4ebb7df9f1120bf3779a1a79a9d5774416f7f78fc34048" Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.708865 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.784864 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-combined-ca-bundle\") pod \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\" (UID: \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\") " Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.784987 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-config-data\") pod \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\" (UID: \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\") " Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.785095 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-logs\") pod \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\" (UID: \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\") " Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.785140 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-87qrq\" (UniqueName: \"kubernetes.io/projected/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-kube-api-access-87qrq\") pod \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\" (UID: \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\") " Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.785188 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-nova-metadata-tls-certs\") pod \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\" (UID: \"4b69ffd2-16ba-45af-b3c9-d4f8b3effebe\") " Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.787484 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-logs" (OuterVolumeSpecName: "logs") pod "4b69ffd2-16ba-45af-b3c9-d4f8b3effebe" (UID: "4b69ffd2-16ba-45af-b3c9-d4f8b3effebe"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.827263 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-kube-api-access-87qrq" (OuterVolumeSpecName: "kube-api-access-87qrq") pod "4b69ffd2-16ba-45af-b3c9-d4f8b3effebe" (UID: "4b69ffd2-16ba-45af-b3c9-d4f8b3effebe"). InnerVolumeSpecName "kube-api-access-87qrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.832253 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-config-data" (OuterVolumeSpecName: "config-data") pod "4b69ffd2-16ba-45af-b3c9-d4f8b3effebe" (UID: "4b69ffd2-16ba-45af-b3c9-d4f8b3effebe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.839462 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4b69ffd2-16ba-45af-b3c9-d4f8b3effebe" (UID: "4b69ffd2-16ba-45af-b3c9-d4f8b3effebe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.876168 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "4b69ffd2-16ba-45af-b3c9-d4f8b3effebe" (UID: "4b69ffd2-16ba-45af-b3c9-d4f8b3effebe"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.889123 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-87qrq\" (UniqueName: \"kubernetes.io/projected/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-kube-api-access-87qrq\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.889157 4813 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.889167 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.889179 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:02 crc kubenswrapper[4813]: I0317 09:34:02.889194 4813 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe-logs\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.694179 4813 generic.go:334] "Generic (PLEG): container finished" podID="69e2e92a-5161-4a09-b59d-464cb4490ff0" containerID="0b9a534ff3bf6cd920494816e2d55e49a081e9a72e266ef15de661be98e603b2" exitCode=0 Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.694243 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"69e2e92a-5161-4a09-b59d-464cb4490ff0","Type":"ContainerDied","Data":"0b9a534ff3bf6cd920494816e2d55e49a081e9a72e266ef15de661be98e603b2"} Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.694553 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.694560 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"69e2e92a-5161-4a09-b59d-464cb4490ff0","Type":"ContainerDied","Data":"a5071088567497d0cc96a3626642895fbd9a3c8d81bd904ae0ad1023200d2336"} Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.694584 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a5071088567497d0cc96a3626642895fbd9a3c8d81bd904ae0ad1023200d2336" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.743150 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.769622 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.784769 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.803854 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Mar 17 09:34:03 crc kubenswrapper[4813]: E0317 09:34:03.804301 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b69ffd2-16ba-45af-b3c9-d4f8b3effebe" containerName="nova-metadata-log" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.804324 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b69ffd2-16ba-45af-b3c9-d4f8b3effebe" containerName="nova-metadata-log" Mar 17 09:34:03 crc kubenswrapper[4813]: E0317 09:34:03.804354 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4b69ffd2-16ba-45af-b3c9-d4f8b3effebe" containerName="nova-metadata-metadata" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.804366 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4b69ffd2-16ba-45af-b3c9-d4f8b3effebe" containerName="nova-metadata-metadata" Mar 17 09:34:03 crc kubenswrapper[4813]: E0317 09:34:03.804380 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69e2e92a-5161-4a09-b59d-464cb4490ff0" containerName="nova-scheduler-scheduler" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.804389 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="69e2e92a-5161-4a09-b59d-464cb4490ff0" containerName="nova-scheduler-scheduler" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.804723 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="69e2e92a-5161-4a09-b59d-464cb4490ff0" containerName="nova-scheduler-scheduler" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.804757 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b69ffd2-16ba-45af-b3c9-d4f8b3effebe" containerName="nova-metadata-log" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.804778 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="4b69ffd2-16ba-45af-b3c9-d4f8b3effebe" containerName="nova-metadata-metadata" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.806029 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.809260 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69e2e92a-5161-4a09-b59d-464cb4490ff0-config-data\") pod \"69e2e92a-5161-4a09-b59d-464cb4490ff0\" (UID: \"69e2e92a-5161-4a09-b59d-464cb4490ff0\") " Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.809343 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfphb\" (UniqueName: \"kubernetes.io/projected/69e2e92a-5161-4a09-b59d-464cb4490ff0-kube-api-access-lfphb\") pod \"69e2e92a-5161-4a09-b59d-464cb4490ff0\" (UID: \"69e2e92a-5161-4a09-b59d-464cb4490ff0\") " Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.809525 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69e2e92a-5161-4a09-b59d-464cb4490ff0-combined-ca-bundle\") pod \"69e2e92a-5161-4a09-b59d-464cb4490ff0\" (UID: \"69e2e92a-5161-4a09-b59d-464cb4490ff0\") " Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.813272 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.813418 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.820902 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.822003 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69e2e92a-5161-4a09-b59d-464cb4490ff0-kube-api-access-lfphb" (OuterVolumeSpecName: "kube-api-access-lfphb") pod "69e2e92a-5161-4a09-b59d-464cb4490ff0" (UID: "69e2e92a-5161-4a09-b59d-464cb4490ff0"). InnerVolumeSpecName "kube-api-access-lfphb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.853538 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69e2e92a-5161-4a09-b59d-464cb4490ff0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "69e2e92a-5161-4a09-b59d-464cb4490ff0" (UID: "69e2e92a-5161-4a09-b59d-464cb4490ff0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.867225 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69e2e92a-5161-4a09-b59d-464cb4490ff0-config-data" (OuterVolumeSpecName: "config-data") pod "69e2e92a-5161-4a09-b59d-464cb4490ff0" (UID: "69e2e92a-5161-4a09-b59d-464cb4490ff0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.919930 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82bd221d-6ac0-4eb2-b709-3de76c656745-logs\") pod \"nova-metadata-0\" (UID: \"82bd221d-6ac0-4eb2-b709-3de76c656745\") " pod="openstack/nova-metadata-0" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.919983 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn9x4\" (UniqueName: \"kubernetes.io/projected/82bd221d-6ac0-4eb2-b709-3de76c656745-kube-api-access-sn9x4\") pod \"nova-metadata-0\" (UID: \"82bd221d-6ac0-4eb2-b709-3de76c656745\") " pod="openstack/nova-metadata-0" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.920176 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82bd221d-6ac0-4eb2-b709-3de76c656745-config-data\") pod \"nova-metadata-0\" (UID: \"82bd221d-6ac0-4eb2-b709-3de76c656745\") " pod="openstack/nova-metadata-0" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.920260 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82bd221d-6ac0-4eb2-b709-3de76c656745-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"82bd221d-6ac0-4eb2-b709-3de76c656745\") " pod="openstack/nova-metadata-0" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.920308 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82bd221d-6ac0-4eb2-b709-3de76c656745-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"82bd221d-6ac0-4eb2-b709-3de76c656745\") " pod="openstack/nova-metadata-0" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.920421 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69e2e92a-5161-4a09-b59d-464cb4490ff0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.920438 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/69e2e92a-5161-4a09-b59d-464cb4490ff0-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:03 crc kubenswrapper[4813]: I0317 09:34:03.920450 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfphb\" (UniqueName: \"kubernetes.io/projected/69e2e92a-5161-4a09-b59d-464cb4490ff0-kube-api-access-lfphb\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.024293 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82bd221d-6ac0-4eb2-b709-3de76c656745-config-data\") pod \"nova-metadata-0\" (UID: \"82bd221d-6ac0-4eb2-b709-3de76c656745\") " pod="openstack/nova-metadata-0" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.024358 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82bd221d-6ac0-4eb2-b709-3de76c656745-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"82bd221d-6ac0-4eb2-b709-3de76c656745\") " pod="openstack/nova-metadata-0" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.024393 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82bd221d-6ac0-4eb2-b709-3de76c656745-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"82bd221d-6ac0-4eb2-b709-3de76c656745\") " pod="openstack/nova-metadata-0" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.024432 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82bd221d-6ac0-4eb2-b709-3de76c656745-logs\") pod \"nova-metadata-0\" (UID: \"82bd221d-6ac0-4eb2-b709-3de76c656745\") " pod="openstack/nova-metadata-0" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.024454 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn9x4\" (UniqueName: \"kubernetes.io/projected/82bd221d-6ac0-4eb2-b709-3de76c656745-kube-api-access-sn9x4\") pod \"nova-metadata-0\" (UID: \"82bd221d-6ac0-4eb2-b709-3de76c656745\") " pod="openstack/nova-metadata-0" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.025142 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82bd221d-6ac0-4eb2-b709-3de76c656745-logs\") pod \"nova-metadata-0\" (UID: \"82bd221d-6ac0-4eb2-b709-3de76c656745\") " pod="openstack/nova-metadata-0" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.027711 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82bd221d-6ac0-4eb2-b709-3de76c656745-config-data\") pod \"nova-metadata-0\" (UID: \"82bd221d-6ac0-4eb2-b709-3de76c656745\") " pod="openstack/nova-metadata-0" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.028124 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82bd221d-6ac0-4eb2-b709-3de76c656745-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"82bd221d-6ac0-4eb2-b709-3de76c656745\") " pod="openstack/nova-metadata-0" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.028726 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82bd221d-6ac0-4eb2-b709-3de76c656745-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"82bd221d-6ac0-4eb2-b709-3de76c656745\") " pod="openstack/nova-metadata-0" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.043258 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn9x4\" (UniqueName: \"kubernetes.io/projected/82bd221d-6ac0-4eb2-b709-3de76c656745-kube-api-access-sn9x4\") pod \"nova-metadata-0\" (UID: \"82bd221d-6ac0-4eb2-b709-3de76c656745\") " pod="openstack/nova-metadata-0" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.126282 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.133800 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562334-bs752" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.226501 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxz47\" (UniqueName: \"kubernetes.io/projected/d363247c-cdbf-49cc-b332-e87f849d9fee-kube-api-access-jxz47\") pod \"d363247c-cdbf-49cc-b332-e87f849d9fee\" (UID: \"d363247c-cdbf-49cc-b332-e87f849d9fee\") " Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.232237 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d363247c-cdbf-49cc-b332-e87f849d9fee-kube-api-access-jxz47" (OuterVolumeSpecName: "kube-api-access-jxz47") pod "d363247c-cdbf-49cc-b332-e87f849d9fee" (UID: "d363247c-cdbf-49cc-b332-e87f849d9fee"). InnerVolumeSpecName "kube-api-access-jxz47". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.330986 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxz47\" (UniqueName: \"kubernetes.io/projected/d363247c-cdbf-49cc-b332-e87f849d9fee-kube-api-access-jxz47\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.618863 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Mar 17 09:34:04 crc kubenswrapper[4813]: W0317 09:34:04.626873 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod82bd221d_6ac0_4eb2_b709_3de76c656745.slice/crio-33fbe9db51be8ae28e5ad863904c62b80b68d54289ee33f5759b8487795cc6e9 WatchSource:0}: Error finding container 33fbe9db51be8ae28e5ad863904c62b80b68d54289ee33f5759b8487795cc6e9: Status 404 returned error can't find the container with id 33fbe9db51be8ae28e5ad863904c62b80b68d54289ee33f5759b8487795cc6e9 Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.711846 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82bd221d-6ac0-4eb2-b709-3de76c656745","Type":"ContainerStarted","Data":"33fbe9db51be8ae28e5ad863904c62b80b68d54289ee33f5759b8487795cc6e9"} Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.714215 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562334-bs752" event={"ID":"d363247c-cdbf-49cc-b332-e87f849d9fee","Type":"ContainerDied","Data":"01d0511642c2f88c9f7f4e431d1a5c7e391c6ee7bfa8f4fb45a565ef41423641"} Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.714256 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="01d0511642c2f88c9f7f4e431d1a5c7e391c6ee7bfa8f4fb45a565ef41423641" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.714263 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562334-bs752" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.714277 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.758553 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4b69ffd2-16ba-45af-b3c9-d4f8b3effebe" path="/var/lib/kubelet/pods/4b69ffd2-16ba-45af-b3c9-d4f8b3effebe/volumes" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.785577 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.803037 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.814132 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 09:34:04 crc kubenswrapper[4813]: E0317 09:34:04.814715 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d363247c-cdbf-49cc-b332-e87f849d9fee" containerName="oc" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.814738 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d363247c-cdbf-49cc-b332-e87f849d9fee" containerName="oc" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.814979 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d363247c-cdbf-49cc-b332-e87f849d9fee" containerName="oc" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.815797 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.824294 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.831882 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.839877 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxcrt\" (UniqueName: \"kubernetes.io/projected/aa41f41c-74ec-44a3-a913-afd1a44d4d04-kube-api-access-pxcrt\") pod \"nova-scheduler-0\" (UID: \"aa41f41c-74ec-44a3-a913-afd1a44d4d04\") " pod="openstack/nova-scheduler-0" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.839959 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa41f41c-74ec-44a3-a913-afd1a44d4d04-config-data\") pod \"nova-scheduler-0\" (UID: \"aa41f41c-74ec-44a3-a913-afd1a44d4d04\") " pod="openstack/nova-scheduler-0" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.840139 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa41f41c-74ec-44a3-a913-afd1a44d4d04-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"aa41f41c-74ec-44a3-a913-afd1a44d4d04\") " pod="openstack/nova-scheduler-0" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.942911 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa41f41c-74ec-44a3-a913-afd1a44d4d04-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"aa41f41c-74ec-44a3-a913-afd1a44d4d04\") " pod="openstack/nova-scheduler-0" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.943019 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxcrt\" (UniqueName: \"kubernetes.io/projected/aa41f41c-74ec-44a3-a913-afd1a44d4d04-kube-api-access-pxcrt\") pod \"nova-scheduler-0\" (UID: \"aa41f41c-74ec-44a3-a913-afd1a44d4d04\") " pod="openstack/nova-scheduler-0" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.943067 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa41f41c-74ec-44a3-a913-afd1a44d4d04-config-data\") pod \"nova-scheduler-0\" (UID: \"aa41f41c-74ec-44a3-a913-afd1a44d4d04\") " pod="openstack/nova-scheduler-0" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.947854 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/aa41f41c-74ec-44a3-a913-afd1a44d4d04-config-data\") pod \"nova-scheduler-0\" (UID: \"aa41f41c-74ec-44a3-a913-afd1a44d4d04\") " pod="openstack/nova-scheduler-0" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.948528 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/aa41f41c-74ec-44a3-a913-afd1a44d4d04-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"aa41f41c-74ec-44a3-a913-afd1a44d4d04\") " pod="openstack/nova-scheduler-0" Mar 17 09:34:04 crc kubenswrapper[4813]: I0317 09:34:04.965393 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxcrt\" (UniqueName: \"kubernetes.io/projected/aa41f41c-74ec-44a3-a913-afd1a44d4d04-kube-api-access-pxcrt\") pod \"nova-scheduler-0\" (UID: \"aa41f41c-74ec-44a3-a913-afd1a44d4d04\") " pod="openstack/nova-scheduler-0" Mar 17 09:34:05 crc kubenswrapper[4813]: I0317 09:34:05.136711 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Mar 17 09:34:05 crc kubenswrapper[4813]: I0317 09:34:05.212932 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562328-pmv66"] Mar 17 09:34:05 crc kubenswrapper[4813]: I0317 09:34:05.223721 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562328-pmv66"] Mar 17 09:34:05 crc kubenswrapper[4813]: I0317 09:34:05.592722 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Mar 17 09:34:05 crc kubenswrapper[4813]: I0317 09:34:05.725088 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"aa41f41c-74ec-44a3-a913-afd1a44d4d04","Type":"ContainerStarted","Data":"4c8c4fc6ae7928d2d2c2b057de6499bb074cae4ed2a69590a6defcc3f88fe2a5"} Mar 17 09:34:05 crc kubenswrapper[4813]: I0317 09:34:05.727436 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82bd221d-6ac0-4eb2-b709-3de76c656745","Type":"ContainerStarted","Data":"abcf49e4f88e59d103f900b4c79fb9b5c2b9429a6f25d9c29fc99ed2e994839b"} Mar 17 09:34:05 crc kubenswrapper[4813]: I0317 09:34:05.727483 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82bd221d-6ac0-4eb2-b709-3de76c656745","Type":"ContainerStarted","Data":"c473b5b5b396a448f7ab6351ca0e42c87b941ea5762698bd5028631c443231b1"} Mar 17 09:34:06 crc kubenswrapper[4813]: I0317 09:34:06.748415 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="02f74478-75b1-4e8a-b347-d0bc8cc32f4c" path="/var/lib/kubelet/pods/02f74478-75b1-4e8a-b347-d0bc8cc32f4c/volumes" Mar 17 09:34:06 crc kubenswrapper[4813]: I0317 09:34:06.749361 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="69e2e92a-5161-4a09-b59d-464cb4490ff0" path="/var/lib/kubelet/pods/69e2e92a-5161-4a09-b59d-464cb4490ff0/volumes" Mar 17 09:34:06 crc kubenswrapper[4813]: I0317 09:34:06.752140 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"aa41f41c-74ec-44a3-a913-afd1a44d4d04","Type":"ContainerStarted","Data":"b1ff3ea712f23c14742b1f17900b377191c88c07a5c1e83c8701cb03de10c6a1"} Mar 17 09:34:06 crc kubenswrapper[4813]: I0317 09:34:06.785000 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.784978484 podStartE2EDuration="2.784978484s" podCreationTimestamp="2026-03-17 09:34:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:34:06.775332021 +0000 UTC m=+1468.876135550" watchObservedRunningTime="2026-03-17 09:34:06.784978484 +0000 UTC m=+1468.885781993" Mar 17 09:34:06 crc kubenswrapper[4813]: I0317 09:34:06.788871 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.788857975 podStartE2EDuration="3.788857975s" podCreationTimestamp="2026-03-17 09:34:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:34:05.762029286 +0000 UTC m=+1467.862832795" watchObservedRunningTime="2026-03-17 09:34:06.788857975 +0000 UTC m=+1468.889661484" Mar 17 09:34:10 crc kubenswrapper[4813]: I0317 09:34:10.137649 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Mar 17 09:34:10 crc kubenswrapper[4813]: I0317 09:34:10.333037 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 17 09:34:10 crc kubenswrapper[4813]: I0317 09:34:10.333648 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Mar 17 09:34:11 crc kubenswrapper[4813]: I0317 09:34:11.349471 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="04ea2c77-25c5-45fc-a375-cf04b185381c" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.214:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 09:34:11 crc kubenswrapper[4813]: I0317 09:34:11.349475 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="04ea2c77-25c5-45fc-a375-cf04b185381c" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.214:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 09:34:11 crc kubenswrapper[4813]: E0317 09:34:11.496880 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod845fdf7b_a3b0_487b_9bcd_efcc3b2b8513.slice\": RecentStats: unable to find data in memory cache]" Mar 17 09:34:14 crc kubenswrapper[4813]: I0317 09:34:14.126728 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 17 09:34:14 crc kubenswrapper[4813]: I0317 09:34:14.126843 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Mar 17 09:34:15 crc kubenswrapper[4813]: I0317 09:34:15.137346 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Mar 17 09:34:15 crc kubenswrapper[4813]: I0317 09:34:15.142857 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="82bd221d-6ac0-4eb2-b709-3de76c656745" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.216:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Mar 17 09:34:15 crc kubenswrapper[4813]: I0317 09:34:15.142870 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="82bd221d-6ac0-4eb2-b709-3de76c656745" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.216:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 17 09:34:15 crc kubenswrapper[4813]: I0317 09:34:15.173560 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Mar 17 09:34:15 crc kubenswrapper[4813]: I0317 09:34:15.921812 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Mar 17 09:34:18 crc kubenswrapper[4813]: I0317 09:34:18.333113 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 17 09:34:18 crc kubenswrapper[4813]: I0317 09:34:18.333477 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Mar 17 09:34:20 crc kubenswrapper[4813]: I0317 09:34:20.367268 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 17 09:34:20 crc kubenswrapper[4813]: I0317 09:34:20.377645 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Mar 17 09:34:20 crc kubenswrapper[4813]: I0317 09:34:20.379302 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 17 09:34:20 crc kubenswrapper[4813]: I0317 09:34:20.951447 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Mar 17 09:34:21 crc kubenswrapper[4813]: E0317 09:34:21.726804 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod845fdf7b_a3b0_487b_9bcd_efcc3b2b8513.slice\": RecentStats: unable to find data in memory cache]" Mar 17 09:34:22 crc kubenswrapper[4813]: I0317 09:34:22.127027 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 17 09:34:22 crc kubenswrapper[4813]: I0317 09:34:22.127129 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Mar 17 09:34:24 crc kubenswrapper[4813]: I0317 09:34:24.135799 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 17 09:34:24 crc kubenswrapper[4813]: I0317 09:34:24.139106 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Mar 17 09:34:24 crc kubenswrapper[4813]: I0317 09:34:24.146642 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 17 09:34:24 crc kubenswrapper[4813]: I0317 09:34:24.989851 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Mar 17 09:34:26 crc kubenswrapper[4813]: I0317 09:34:26.010029 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Mar 17 09:34:32 crc kubenswrapper[4813]: E0317 09:34:32.012887 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod845fdf7b_a3b0_487b_9bcd_efcc3b2b8513.slice\": RecentStats: unable to find data in memory cache]" Mar 17 09:34:34 crc kubenswrapper[4813]: I0317 09:34:34.920600 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 17 09:34:35 crc kubenswrapper[4813]: I0317 09:34:35.745382 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 17 09:34:38 crc kubenswrapper[4813]: I0317 09:34:38.972969 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="2517b76a-309a-4c9e-88e7-6061d528acc8" containerName="rabbitmq" containerID="cri-o://38f5e639d16fd0e9a3c8edcf8dce798b0c07cc4284524957e965004e5a4ee557" gracePeriod=604796 Mar 17 09:34:39 crc kubenswrapper[4813]: I0317 09:34:39.772904 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="d84c4ab5-7f54-48dd-b45c-03c9dd1ee753" containerName="rabbitmq" containerID="cri-o://202647fbd1a9ea05b2b013ade8131b58b3d231ff0d159880c4bbdc45f0385dc8" gracePeriod=604796 Mar 17 09:34:41 crc kubenswrapper[4813]: I0317 09:34:41.328546 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-w5vxx"] Mar 17 09:34:41 crc kubenswrapper[4813]: I0317 09:34:41.330429 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w5vxx" Mar 17 09:34:41 crc kubenswrapper[4813]: I0317 09:34:41.363695 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w5vxx"] Mar 17 09:34:41 crc kubenswrapper[4813]: I0317 09:34:41.391222 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee189143-6f3d-489d-be41-604ee0f72772-catalog-content\") pod \"redhat-operators-w5vxx\" (UID: \"ee189143-6f3d-489d-be41-604ee0f72772\") " pod="openshift-marketplace/redhat-operators-w5vxx" Mar 17 09:34:41 crc kubenswrapper[4813]: I0317 09:34:41.391366 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee189143-6f3d-489d-be41-604ee0f72772-utilities\") pod \"redhat-operators-w5vxx\" (UID: \"ee189143-6f3d-489d-be41-604ee0f72772\") " pod="openshift-marketplace/redhat-operators-w5vxx" Mar 17 09:34:41 crc kubenswrapper[4813]: I0317 09:34:41.391713 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dkht2\" (UniqueName: \"kubernetes.io/projected/ee189143-6f3d-489d-be41-604ee0f72772-kube-api-access-dkht2\") pod \"redhat-operators-w5vxx\" (UID: \"ee189143-6f3d-489d-be41-604ee0f72772\") " pod="openshift-marketplace/redhat-operators-w5vxx" Mar 17 09:34:41 crc kubenswrapper[4813]: I0317 09:34:41.493111 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dkht2\" (UniqueName: \"kubernetes.io/projected/ee189143-6f3d-489d-be41-604ee0f72772-kube-api-access-dkht2\") pod \"redhat-operators-w5vxx\" (UID: \"ee189143-6f3d-489d-be41-604ee0f72772\") " pod="openshift-marketplace/redhat-operators-w5vxx" Mar 17 09:34:41 crc kubenswrapper[4813]: I0317 09:34:41.493169 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee189143-6f3d-489d-be41-604ee0f72772-catalog-content\") pod \"redhat-operators-w5vxx\" (UID: \"ee189143-6f3d-489d-be41-604ee0f72772\") " pod="openshift-marketplace/redhat-operators-w5vxx" Mar 17 09:34:41 crc kubenswrapper[4813]: I0317 09:34:41.493204 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee189143-6f3d-489d-be41-604ee0f72772-utilities\") pod \"redhat-operators-w5vxx\" (UID: \"ee189143-6f3d-489d-be41-604ee0f72772\") " pod="openshift-marketplace/redhat-operators-w5vxx" Mar 17 09:34:41 crc kubenswrapper[4813]: I0317 09:34:41.493745 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee189143-6f3d-489d-be41-604ee0f72772-utilities\") pod \"redhat-operators-w5vxx\" (UID: \"ee189143-6f3d-489d-be41-604ee0f72772\") " pod="openshift-marketplace/redhat-operators-w5vxx" Mar 17 09:34:41 crc kubenswrapper[4813]: I0317 09:34:41.493891 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee189143-6f3d-489d-be41-604ee0f72772-catalog-content\") pod \"redhat-operators-w5vxx\" (UID: \"ee189143-6f3d-489d-be41-604ee0f72772\") " pod="openshift-marketplace/redhat-operators-w5vxx" Mar 17 09:34:41 crc kubenswrapper[4813]: I0317 09:34:41.528351 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dkht2\" (UniqueName: \"kubernetes.io/projected/ee189143-6f3d-489d-be41-604ee0f72772-kube-api-access-dkht2\") pod \"redhat-operators-w5vxx\" (UID: \"ee189143-6f3d-489d-be41-604ee0f72772\") " pod="openshift-marketplace/redhat-operators-w5vxx" Mar 17 09:34:41 crc kubenswrapper[4813]: I0317 09:34:41.665831 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w5vxx" Mar 17 09:34:42 crc kubenswrapper[4813]: W0317 09:34:42.184001 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podee189143_6f3d_489d_be41_604ee0f72772.slice/crio-92deef0c321d544683d9b8e7d8cf3452850f0826efa913fc47004b16e5d62156 WatchSource:0}: Error finding container 92deef0c321d544683d9b8e7d8cf3452850f0826efa913fc47004b16e5d62156: Status 404 returned error can't find the container with id 92deef0c321d544683d9b8e7d8cf3452850f0826efa913fc47004b16e5d62156 Mar 17 09:34:42 crc kubenswrapper[4813]: I0317 09:34:42.187512 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-w5vxx"] Mar 17 09:34:43 crc kubenswrapper[4813]: I0317 09:34:43.153897 4813 generic.go:334] "Generic (PLEG): container finished" podID="ee189143-6f3d-489d-be41-604ee0f72772" containerID="ed7212d277e75ed05c93691ac8c9504193a643c1d280b36a30e51ce6a0b7e14a" exitCode=0 Mar 17 09:34:43 crc kubenswrapper[4813]: I0317 09:34:43.153965 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5vxx" event={"ID":"ee189143-6f3d-489d-be41-604ee0f72772","Type":"ContainerDied","Data":"ed7212d277e75ed05c93691ac8c9504193a643c1d280b36a30e51ce6a0b7e14a"} Mar 17 09:34:43 crc kubenswrapper[4813]: I0317 09:34:43.154211 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5vxx" event={"ID":"ee189143-6f3d-489d-be41-604ee0f72772","Type":"ContainerStarted","Data":"92deef0c321d544683d9b8e7d8cf3452850f0826efa913fc47004b16e5d62156"} Mar 17 09:34:44 crc kubenswrapper[4813]: I0317 09:34:44.165839 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5vxx" event={"ID":"ee189143-6f3d-489d-be41-604ee0f72772","Type":"ContainerStarted","Data":"a711b4fe15101b0fc9a50d6790cb354c4d562f0c35a172cee29b51958a8195a6"} Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.232962 4813 scope.go:117] "RemoveContainer" containerID="f60ccc154e2565faf3d3ff26f1d684ed650435d011b99faac17cdf34034d77c9" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.481993 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-8pbz7"] Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.486858 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8pbz7" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.490214 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8pbz7"] Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.585387 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1116d60a-9394-4311-ab91-fa9181c8dd66-utilities\") pod \"community-operators-8pbz7\" (UID: \"1116d60a-9394-4311-ab91-fa9181c8dd66\") " pod="openshift-marketplace/community-operators-8pbz7" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.585509 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1116d60a-9394-4311-ab91-fa9181c8dd66-catalog-content\") pod \"community-operators-8pbz7\" (UID: \"1116d60a-9394-4311-ab91-fa9181c8dd66\") " pod="openshift-marketplace/community-operators-8pbz7" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.585552 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mn8sp\" (UniqueName: \"kubernetes.io/projected/1116d60a-9394-4311-ab91-fa9181c8dd66-kube-api-access-mn8sp\") pod \"community-operators-8pbz7\" (UID: \"1116d60a-9394-4311-ab91-fa9181c8dd66\") " pod="openshift-marketplace/community-operators-8pbz7" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.598711 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.693247 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2517b76a-309a-4c9e-88e7-6061d528acc8-plugins-conf\") pod \"2517b76a-309a-4c9e-88e7-6061d528acc8\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.693321 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2517b76a-309a-4c9e-88e7-6061d528acc8-pod-info\") pod \"2517b76a-309a-4c9e-88e7-6061d528acc8\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.693370 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-plugins\") pod \"2517b76a-309a-4c9e-88e7-6061d528acc8\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.693468 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-erlang-cookie\") pod \"2517b76a-309a-4c9e-88e7-6061d528acc8\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.693503 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2517b76a-309a-4c9e-88e7-6061d528acc8-config-data\") pod \"2517b76a-309a-4c9e-88e7-6061d528acc8\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.693552 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2517b76a-309a-4c9e-88e7-6061d528acc8-erlang-cookie-secret\") pod \"2517b76a-309a-4c9e-88e7-6061d528acc8\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.693590 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-tls\") pod \"2517b76a-309a-4c9e-88e7-6061d528acc8\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.693639 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2517b76a-309a-4c9e-88e7-6061d528acc8-server-conf\") pod \"2517b76a-309a-4c9e-88e7-6061d528acc8\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.693682 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-confd\") pod \"2517b76a-309a-4c9e-88e7-6061d528acc8\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.693736 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcnpw\" (UniqueName: \"kubernetes.io/projected/2517b76a-309a-4c9e-88e7-6061d528acc8-kube-api-access-rcnpw\") pod \"2517b76a-309a-4c9e-88e7-6061d528acc8\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.693762 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"2517b76a-309a-4c9e-88e7-6061d528acc8\" (UID: \"2517b76a-309a-4c9e-88e7-6061d528acc8\") " Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.694002 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1116d60a-9394-4311-ab91-fa9181c8dd66-utilities\") pod \"community-operators-8pbz7\" (UID: \"1116d60a-9394-4311-ab91-fa9181c8dd66\") " pod="openshift-marketplace/community-operators-8pbz7" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.694091 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1116d60a-9394-4311-ab91-fa9181c8dd66-catalog-content\") pod \"community-operators-8pbz7\" (UID: \"1116d60a-9394-4311-ab91-fa9181c8dd66\") " pod="openshift-marketplace/community-operators-8pbz7" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.694130 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mn8sp\" (UniqueName: \"kubernetes.io/projected/1116d60a-9394-4311-ab91-fa9181c8dd66-kube-api-access-mn8sp\") pod \"community-operators-8pbz7\" (UID: \"1116d60a-9394-4311-ab91-fa9181c8dd66\") " pod="openshift-marketplace/community-operators-8pbz7" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.694194 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2517b76a-309a-4c9e-88e7-6061d528acc8-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "2517b76a-309a-4c9e-88e7-6061d528acc8" (UID: "2517b76a-309a-4c9e-88e7-6061d528acc8"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.699238 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "2517b76a-309a-4c9e-88e7-6061d528acc8" (UID: "2517b76a-309a-4c9e-88e7-6061d528acc8"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.700332 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "persistence") pod "2517b76a-309a-4c9e-88e7-6061d528acc8" (UID: "2517b76a-309a-4c9e-88e7-6061d528acc8"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.700957 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "2517b76a-309a-4c9e-88e7-6061d528acc8" (UID: "2517b76a-309a-4c9e-88e7-6061d528acc8"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.701544 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "2517b76a-309a-4c9e-88e7-6061d528acc8" (UID: "2517b76a-309a-4c9e-88e7-6061d528acc8"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.701889 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1116d60a-9394-4311-ab91-fa9181c8dd66-catalog-content\") pod \"community-operators-8pbz7\" (UID: \"1116d60a-9394-4311-ab91-fa9181c8dd66\") " pod="openshift-marketplace/community-operators-8pbz7" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.702381 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1116d60a-9394-4311-ab91-fa9181c8dd66-utilities\") pod \"community-operators-8pbz7\" (UID: \"1116d60a-9394-4311-ab91-fa9181c8dd66\") " pod="openshift-marketplace/community-operators-8pbz7" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.703290 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2517b76a-309a-4c9e-88e7-6061d528acc8-kube-api-access-rcnpw" (OuterVolumeSpecName: "kube-api-access-rcnpw") pod "2517b76a-309a-4c9e-88e7-6061d528acc8" (UID: "2517b76a-309a-4c9e-88e7-6061d528acc8"). InnerVolumeSpecName "kube-api-access-rcnpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.709520 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/2517b76a-309a-4c9e-88e7-6061d528acc8-pod-info" (OuterVolumeSpecName: "pod-info") pod "2517b76a-309a-4c9e-88e7-6061d528acc8" (UID: "2517b76a-309a-4c9e-88e7-6061d528acc8"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.709799 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2517b76a-309a-4c9e-88e7-6061d528acc8-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "2517b76a-309a-4c9e-88e7-6061d528acc8" (UID: "2517b76a-309a-4c9e-88e7-6061d528acc8"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.724909 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mn8sp\" (UniqueName: \"kubernetes.io/projected/1116d60a-9394-4311-ab91-fa9181c8dd66-kube-api-access-mn8sp\") pod \"community-operators-8pbz7\" (UID: \"1116d60a-9394-4311-ab91-fa9181c8dd66\") " pod="openshift-marketplace/community-operators-8pbz7" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.742493 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2517b76a-309a-4c9e-88e7-6061d528acc8-config-data" (OuterVolumeSpecName: "config-data") pod "2517b76a-309a-4c9e-88e7-6061d528acc8" (UID: "2517b76a-309a-4c9e-88e7-6061d528acc8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.773254 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2517b76a-309a-4c9e-88e7-6061d528acc8-server-conf" (OuterVolumeSpecName: "server-conf") pod "2517b76a-309a-4c9e-88e7-6061d528acc8" (UID: "2517b76a-309a-4c9e-88e7-6061d528acc8"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.796515 4813 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.796552 4813 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/2517b76a-309a-4c9e-88e7-6061d528acc8-server-conf\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.796565 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcnpw\" (UniqueName: \"kubernetes.io/projected/2517b76a-309a-4c9e-88e7-6061d528acc8-kube-api-access-rcnpw\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.796597 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.796640 4813 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/2517b76a-309a-4c9e-88e7-6061d528acc8-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.796652 4813 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/2517b76a-309a-4c9e-88e7-6061d528acc8-pod-info\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.796662 4813 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.796672 4813 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.796682 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/2517b76a-309a-4c9e-88e7-6061d528acc8-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.796691 4813 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/2517b76a-309a-4c9e-88e7-6061d528acc8-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.808033 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8pbz7" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.818135 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.845912 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "2517b76a-309a-4c9e-88e7-6061d528acc8" (UID: "2517b76a-309a-4c9e-88e7-6061d528acc8"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.898512 4813 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/2517b76a-309a-4c9e-88e7-6061d528acc8-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:45 crc kubenswrapper[4813]: I0317 09:34:45.898543 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.046008 4813 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="d84c4ab5-7f54-48dd-b45c-03c9dd1ee753" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.214522 4813 generic.go:334] "Generic (PLEG): container finished" podID="d84c4ab5-7f54-48dd-b45c-03c9dd1ee753" containerID="202647fbd1a9ea05b2b013ade8131b58b3d231ff0d159880c4bbdc45f0385dc8" exitCode=0 Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.214616 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753","Type":"ContainerDied","Data":"202647fbd1a9ea05b2b013ade8131b58b3d231ff0d159880c4bbdc45f0385dc8"} Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.216965 4813 generic.go:334] "Generic (PLEG): container finished" podID="2517b76a-309a-4c9e-88e7-6061d528acc8" containerID="38f5e639d16fd0e9a3c8edcf8dce798b0c07cc4284524957e965004e5a4ee557" exitCode=0 Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.217065 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2517b76a-309a-4c9e-88e7-6061d528acc8","Type":"ContainerDied","Data":"38f5e639d16fd0e9a3c8edcf8dce798b0c07cc4284524957e965004e5a4ee557"} Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.217133 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"2517b76a-309a-4c9e-88e7-6061d528acc8","Type":"ContainerDied","Data":"f07311e221541fb35044925dc0c83c28c25c05e7d66b76a0cb68aef674ef228f"} Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.217157 4813 scope.go:117] "RemoveContainer" containerID="38f5e639d16fd0e9a3c8edcf8dce798b0c07cc4284524957e965004e5a4ee557" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.217190 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.271236 4813 scope.go:117] "RemoveContainer" containerID="d5713bfe8ab956c6c4a2dacaf5fc3b1e6cb4739a7e1e69e2bbd95c16f91dc7df" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.280125 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.300145 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-8pbz7"] Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.316882 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.334723 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Mar 17 09:34:46 crc kubenswrapper[4813]: E0317 09:34:46.335351 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2517b76a-309a-4c9e-88e7-6061d528acc8" containerName="setup-container" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.335389 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="2517b76a-309a-4c9e-88e7-6061d528acc8" containerName="setup-container" Mar 17 09:34:46 crc kubenswrapper[4813]: E0317 09:34:46.335419 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2517b76a-309a-4c9e-88e7-6061d528acc8" containerName="rabbitmq" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.335426 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="2517b76a-309a-4c9e-88e7-6061d528acc8" containerName="rabbitmq" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.335710 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="2517b76a-309a-4c9e-88e7-6061d528acc8" containerName="rabbitmq" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.336956 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.352004 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.354186 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.354405 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.354514 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-xfkfx" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.354710 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.354821 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.354927 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.355038 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.436904 4813 scope.go:117] "RemoveContainer" containerID="38f5e639d16fd0e9a3c8edcf8dce798b0c07cc4284524957e965004e5a4ee557" Mar 17 09:34:46 crc kubenswrapper[4813]: E0317 09:34:46.446209 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38f5e639d16fd0e9a3c8edcf8dce798b0c07cc4284524957e965004e5a4ee557\": container with ID starting with 38f5e639d16fd0e9a3c8edcf8dce798b0c07cc4284524957e965004e5a4ee557 not found: ID does not exist" containerID="38f5e639d16fd0e9a3c8edcf8dce798b0c07cc4284524957e965004e5a4ee557" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.446256 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38f5e639d16fd0e9a3c8edcf8dce798b0c07cc4284524957e965004e5a4ee557"} err="failed to get container status \"38f5e639d16fd0e9a3c8edcf8dce798b0c07cc4284524957e965004e5a4ee557\": rpc error: code = NotFound desc = could not find container \"38f5e639d16fd0e9a3c8edcf8dce798b0c07cc4284524957e965004e5a4ee557\": container with ID starting with 38f5e639d16fd0e9a3c8edcf8dce798b0c07cc4284524957e965004e5a4ee557 not found: ID does not exist" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.446279 4813 scope.go:117] "RemoveContainer" containerID="d5713bfe8ab956c6c4a2dacaf5fc3b1e6cb4739a7e1e69e2bbd95c16f91dc7df" Mar 17 09:34:46 crc kubenswrapper[4813]: E0317 09:34:46.464782 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5713bfe8ab956c6c4a2dacaf5fc3b1e6cb4739a7e1e69e2bbd95c16f91dc7df\": container with ID starting with d5713bfe8ab956c6c4a2dacaf5fc3b1e6cb4739a7e1e69e2bbd95c16f91dc7df not found: ID does not exist" containerID="d5713bfe8ab956c6c4a2dacaf5fc3b1e6cb4739a7e1e69e2bbd95c16f91dc7df" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.464827 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5713bfe8ab956c6c4a2dacaf5fc3b1e6cb4739a7e1e69e2bbd95c16f91dc7df"} err="failed to get container status \"d5713bfe8ab956c6c4a2dacaf5fc3b1e6cb4739a7e1e69e2bbd95c16f91dc7df\": rpc error: code = NotFound desc = could not find container \"d5713bfe8ab956c6c4a2dacaf5fc3b1e6cb4739a7e1e69e2bbd95c16f91dc7df\": container with ID starting with d5713bfe8ab956c6c4a2dacaf5fc3b1e6cb4739a7e1e69e2bbd95c16f91dc7df not found: ID does not exist" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.512179 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.512244 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.512265 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.512288 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.512307 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kz624\" (UniqueName: \"kubernetes.io/projected/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-kube-api-access-kz624\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.512356 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.512404 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.512423 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.512444 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-config-data\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.512475 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.512494 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.614276 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.614322 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.614346 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-config-data\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.614376 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.614391 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.614424 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.614467 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.614483 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.614506 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.614524 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kz624\" (UniqueName: \"kubernetes.io/projected/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-kube-api-access-kz624\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.614568 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.615512 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.616274 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.619310 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.619566 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-config-data\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.620326 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.622445 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.622993 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.625812 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.626291 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.640491 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.645176 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kz624\" (UniqueName: \"kubernetes.io/projected/16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b-kube-api-access-kz624\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.654392 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"rabbitmq-server-0\" (UID: \"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b\") " pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.678123 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.722234 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.744093 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2517b76a-309a-4c9e-88e7-6061d528acc8" path="/var/lib/kubelet/pods/2517b76a-309a-4c9e-88e7-6061d528acc8/volumes" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.918995 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-erlang-cookie-secret\") pod \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.919054 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-erlang-cookie\") pod \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.919119 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m28ms\" (UniqueName: \"kubernetes.io/projected/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-kube-api-access-m28ms\") pod \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.919175 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-confd\") pod \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.919229 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-config-data\") pod \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.919351 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-pod-info\") pod \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.919419 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-plugins\") pod \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.919467 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-server-conf\") pod \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.919495 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-tls\") pod \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.919525 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-plugins-conf\") pod \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.919556 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\" (UID: \"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753\") " Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.920136 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753" (UID: "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.920568 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753" (UID: "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.920714 4813 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.922094 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753" (UID: "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.923731 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753" (UID: "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.924367 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-pod-info" (OuterVolumeSpecName: "pod-info") pod "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753" (UID: "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.925783 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "persistence") pod "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753" (UID: "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.926189 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-kube-api-access-m28ms" (OuterVolumeSpecName: "kube-api-access-m28ms") pod "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753" (UID: "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753"). InnerVolumeSpecName "kube-api-access-m28ms". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.927894 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753" (UID: "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.954173 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-config-data" (OuterVolumeSpecName: "config-data") pod "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753" (UID: "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:34:46 crc kubenswrapper[4813]: I0317 09:34:46.984544 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-server-conf" (OuterVolumeSpecName: "server-conf") pod "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753" (UID: "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.023282 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m28ms\" (UniqueName: \"kubernetes.io/projected/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-kube-api-access-m28ms\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.023327 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.023337 4813 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-pod-info\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.023345 4813 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.023353 4813 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-server-conf\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.023363 4813 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.023371 4813 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-plugins-conf\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.023413 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.023424 4813 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.028909 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753" (UID: "d84c4ab5-7f54-48dd-b45c-03c9dd1ee753"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.047826 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.124881 4813 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.124916 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.229679 4813 generic.go:334] "Generic (PLEG): container finished" podID="1116d60a-9394-4311-ab91-fa9181c8dd66" containerID="5cff32a263fac1ab39538939589f2a42dd0a96beaadfb8e9e90b22218998819b" exitCode=0 Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.229749 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8pbz7" event={"ID":"1116d60a-9394-4311-ab91-fa9181c8dd66","Type":"ContainerDied","Data":"5cff32a263fac1ab39538939589f2a42dd0a96beaadfb8e9e90b22218998819b"} Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.229776 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8pbz7" event={"ID":"1116d60a-9394-4311-ab91-fa9181c8dd66","Type":"ContainerStarted","Data":"32f4aa866b8d9fbe75e9df5419efc5ba185675147f808adb6ea4cc087e2fee72"} Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.234104 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"d84c4ab5-7f54-48dd-b45c-03c9dd1ee753","Type":"ContainerDied","Data":"f419aa470ce9c02a701fd2d50b2ce9af2a29a9d337962090cd3f469c436d39af"} Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.234140 4813 scope.go:117] "RemoveContainer" containerID="202647fbd1a9ea05b2b013ade8131b58b3d231ff0d159880c4bbdc45f0385dc8" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.234257 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.280258 4813 scope.go:117] "RemoveContainer" containerID="92494a89ef4e16780b97fced28f929444d04138f37ecab69d9f1fcef95b6682f" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.297106 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.318644 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.328978 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.347716 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 17 09:34:47 crc kubenswrapper[4813]: E0317 09:34:47.348428 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d84c4ab5-7f54-48dd-b45c-03c9dd1ee753" containerName="rabbitmq" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.348440 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d84c4ab5-7f54-48dd-b45c-03c9dd1ee753" containerName="rabbitmq" Mar 17 09:34:47 crc kubenswrapper[4813]: E0317 09:34:47.348453 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d84c4ab5-7f54-48dd-b45c-03c9dd1ee753" containerName="setup-container" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.348459 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d84c4ab5-7f54-48dd-b45c-03c9dd1ee753" containerName="setup-container" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.348661 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d84c4ab5-7f54-48dd-b45c-03c9dd1ee753" containerName="rabbitmq" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.350397 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.356778 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-5rdjh" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.356900 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.356958 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.357118 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.357183 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.357239 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.357236 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.361316 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.555680 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cfc379d0-d53a-4640-8d3f-a883db1aefa9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.555722 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.555760 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cfc379d0-d53a-4640-8d3f-a883db1aefa9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.555778 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cfc379d0-d53a-4640-8d3f-a883db1aefa9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.555802 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cfc379d0-d53a-4640-8d3f-a883db1aefa9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.555943 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cfc379d0-d53a-4640-8d3f-a883db1aefa9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.556026 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cfc379d0-d53a-4640-8d3f-a883db1aefa9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.556169 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cfc379d0-d53a-4640-8d3f-a883db1aefa9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.556229 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cfc379d0-d53a-4640-8d3f-a883db1aefa9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.556391 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cfc379d0-d53a-4640-8d3f-a883db1aefa9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.556513 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqztw\" (UniqueName: \"kubernetes.io/projected/cfc379d0-d53a-4640-8d3f-a883db1aefa9-kube-api-access-qqztw\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.658518 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cfc379d0-d53a-4640-8d3f-a883db1aefa9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.658559 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cfc379d0-d53a-4640-8d3f-a883db1aefa9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.658587 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cfc379d0-d53a-4640-8d3f-a883db1aefa9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.658644 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cfc379d0-d53a-4640-8d3f-a883db1aefa9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.659205 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/cfc379d0-d53a-4640-8d3f-a883db1aefa9-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.658665 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cfc379d0-d53a-4640-8d3f-a883db1aefa9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.659511 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cfc379d0-d53a-4640-8d3f-a883db1aefa9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.659571 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cfc379d0-d53a-4640-8d3f-a883db1aefa9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.659615 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/cfc379d0-d53a-4640-8d3f-a883db1aefa9-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.659712 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cfc379d0-d53a-4640-8d3f-a883db1aefa9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.659798 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqztw\" (UniqueName: \"kubernetes.io/projected/cfc379d0-d53a-4640-8d3f-a883db1aefa9-kube-api-access-qqztw\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.659914 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cfc379d0-d53a-4640-8d3f-a883db1aefa9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.659950 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.660248 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.660917 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/cfc379d0-d53a-4640-8d3f-a883db1aefa9-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.661034 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/cfc379d0-d53a-4640-8d3f-a883db1aefa9-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.662107 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/cfc379d0-d53a-4640-8d3f-a883db1aefa9-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.663274 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/cfc379d0-d53a-4640-8d3f-a883db1aefa9-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.663767 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/cfc379d0-d53a-4640-8d3f-a883db1aefa9-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.663775 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/cfc379d0-d53a-4640-8d3f-a883db1aefa9-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.664258 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/cfc379d0-d53a-4640-8d3f-a883db1aefa9-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.676535 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqztw\" (UniqueName: \"kubernetes.io/projected/cfc379d0-d53a-4640-8d3f-a883db1aefa9-kube-api-access-qqztw\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.689306 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"cfc379d0-d53a-4640-8d3f-a883db1aefa9\") " pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.731457 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.943000 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-r46bt"] Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.966579 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.971800 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Mar 17 09:34:47 crc kubenswrapper[4813]: I0317 09:34:47.978121 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-r46bt"] Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.006429 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-r46bt\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.006481 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-r46bt\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.006544 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-r46bt\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.006584 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5f5mz\" (UniqueName: \"kubernetes.io/projected/af669e24-da1e-4585-865c-e6d978c12838-kube-api-access-5f5mz\") pod \"dnsmasq-dns-67b789f86c-r46bt\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.006638 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-dns-svc\") pod \"dnsmasq-dns-67b789f86c-r46bt\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.006668 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-config\") pod \"dnsmasq-dns-67b789f86c-r46bt\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.006710 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-r46bt\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.109747 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-r46bt\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.110090 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-r46bt\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.110143 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-r46bt\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.110175 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5f5mz\" (UniqueName: \"kubernetes.io/projected/af669e24-da1e-4585-865c-e6d978c12838-kube-api-access-5f5mz\") pod \"dnsmasq-dns-67b789f86c-r46bt\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.110202 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-dns-svc\") pod \"dnsmasq-dns-67b789f86c-r46bt\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.110233 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-config\") pod \"dnsmasq-dns-67b789f86c-r46bt\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.110266 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-r46bt\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.111067 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-ovsdbserver-nb\") pod \"dnsmasq-dns-67b789f86c-r46bt\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.111237 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-openstack-edpm-ipam\") pod \"dnsmasq-dns-67b789f86c-r46bt\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.111437 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-dns-swift-storage-0\") pod \"dnsmasq-dns-67b789f86c-r46bt\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.112016 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-dns-svc\") pod \"dnsmasq-dns-67b789f86c-r46bt\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.112034 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-ovsdbserver-sb\") pod \"dnsmasq-dns-67b789f86c-r46bt\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.113024 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-config\") pod \"dnsmasq-dns-67b789f86c-r46bt\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.133728 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5f5mz\" (UniqueName: \"kubernetes.io/projected/af669e24-da1e-4585-865c-e6d978c12838-kube-api-access-5f5mz\") pod \"dnsmasq-dns-67b789f86c-r46bt\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: W0317 09:34:48.213874 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfc379d0_d53a_4640_8d3f_a883db1aefa9.slice/crio-601390fc16f53f6f062902349987f9d6db1378dcd10e3625eef3bb6e2b1e2936 WatchSource:0}: Error finding container 601390fc16f53f6f062902349987f9d6db1378dcd10e3625eef3bb6e2b1e2936: Status 404 returned error can't find the container with id 601390fc16f53f6f062902349987f9d6db1378dcd10e3625eef3bb6e2b1e2936 Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.215149 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.244814 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cfc379d0-d53a-4640-8d3f-a883db1aefa9","Type":"ContainerStarted","Data":"601390fc16f53f6f062902349987f9d6db1378dcd10e3625eef3bb6e2b1e2936"} Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.246552 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8pbz7" event={"ID":"1116d60a-9394-4311-ab91-fa9181c8dd66","Type":"ContainerStarted","Data":"a5b930d6471e1aa9daf3e525cc453d1d99d6129b90d6d09f93265174de9ddc2a"} Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.248281 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b","Type":"ContainerStarted","Data":"bb9ca7f67079c50e6d3d8dd73f0574034c838ec9b38918622d3fc4f69bcd6bb0"} Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.335539 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.746201 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d84c4ab5-7f54-48dd-b45c-03c9dd1ee753" path="/var/lib/kubelet/pods/d84c4ab5-7f54-48dd-b45c-03c9dd1ee753/volumes" Mar 17 09:34:48 crc kubenswrapper[4813]: W0317 09:34:48.818818 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf669e24_da1e_4585_865c_e6d978c12838.slice/crio-635abe893adfd1281af877443424901a91f190c1edac8822b91241e09224c847 WatchSource:0}: Error finding container 635abe893adfd1281af877443424901a91f190c1edac8822b91241e09224c847: Status 404 returned error can't find the container with id 635abe893adfd1281af877443424901a91f190c1edac8822b91241e09224c847 Mar 17 09:34:48 crc kubenswrapper[4813]: I0317 09:34:48.823013 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-r46bt"] Mar 17 09:34:49 crc kubenswrapper[4813]: I0317 09:34:49.263937 4813 generic.go:334] "Generic (PLEG): container finished" podID="1116d60a-9394-4311-ab91-fa9181c8dd66" containerID="a5b930d6471e1aa9daf3e525cc453d1d99d6129b90d6d09f93265174de9ddc2a" exitCode=0 Mar 17 09:34:49 crc kubenswrapper[4813]: I0317 09:34:49.264118 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8pbz7" event={"ID":"1116d60a-9394-4311-ab91-fa9181c8dd66","Type":"ContainerDied","Data":"a5b930d6471e1aa9daf3e525cc453d1d99d6129b90d6d09f93265174de9ddc2a"} Mar 17 09:34:49 crc kubenswrapper[4813]: I0317 09:34:49.266122 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b","Type":"ContainerStarted","Data":"a1e83709dd2273fea789e5921d2b30f17f4925695ba5ce7811df41ecec5bd30c"} Mar 17 09:34:49 crc kubenswrapper[4813]: I0317 09:34:49.269638 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-r46bt" event={"ID":"af669e24-da1e-4585-865c-e6d978c12838","Type":"ContainerDied","Data":"fc487a06c5503d88b64ffd200cae2436b83e8922ce174dff8d881b7437ea83c0"} Mar 17 09:34:49 crc kubenswrapper[4813]: I0317 09:34:49.269576 4813 generic.go:334] "Generic (PLEG): container finished" podID="af669e24-da1e-4585-865c-e6d978c12838" containerID="fc487a06c5503d88b64ffd200cae2436b83e8922ce174dff8d881b7437ea83c0" exitCode=0 Mar 17 09:34:49 crc kubenswrapper[4813]: I0317 09:34:49.269745 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-r46bt" event={"ID":"af669e24-da1e-4585-865c-e6d978c12838","Type":"ContainerStarted","Data":"635abe893adfd1281af877443424901a91f190c1edac8822b91241e09224c847"} Mar 17 09:34:50 crc kubenswrapper[4813]: I0317 09:34:50.280207 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8pbz7" event={"ID":"1116d60a-9394-4311-ab91-fa9181c8dd66","Type":"ContainerStarted","Data":"6329b5e6bfca5977e464952a606bf8051ce46982f04a378ac87acff18c020c9a"} Mar 17 09:34:50 crc kubenswrapper[4813]: I0317 09:34:50.282747 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-r46bt" event={"ID":"af669e24-da1e-4585-865c-e6d978c12838","Type":"ContainerStarted","Data":"f95998cbf5bb06f0793dbfffb2b455c0da8571e30177362a52d817b3aefc91d8"} Mar 17 09:34:50 crc kubenswrapper[4813]: I0317 09:34:50.282872 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:50 crc kubenswrapper[4813]: I0317 09:34:50.284383 4813 generic.go:334] "Generic (PLEG): container finished" podID="ee189143-6f3d-489d-be41-604ee0f72772" containerID="a711b4fe15101b0fc9a50d6790cb354c4d562f0c35a172cee29b51958a8195a6" exitCode=0 Mar 17 09:34:50 crc kubenswrapper[4813]: I0317 09:34:50.284436 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5vxx" event={"ID":"ee189143-6f3d-489d-be41-604ee0f72772","Type":"ContainerDied","Data":"a711b4fe15101b0fc9a50d6790cb354c4d562f0c35a172cee29b51958a8195a6"} Mar 17 09:34:50 crc kubenswrapper[4813]: I0317 09:34:50.286086 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cfc379d0-d53a-4640-8d3f-a883db1aefa9","Type":"ContainerStarted","Data":"e6db8e4fb9e477f6fd0d51a9ab4a9b4a5f2770560a5526c186c89b4622d1afdd"} Mar 17 09:34:50 crc kubenswrapper[4813]: I0317 09:34:50.300685 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-8pbz7" podStartSLOduration=2.826526905 podStartE2EDuration="5.300659414s" podCreationTimestamp="2026-03-17 09:34:45 +0000 UTC" firstStartedPulling="2026-03-17 09:34:47.231985035 +0000 UTC m=+1509.332788534" lastFinishedPulling="2026-03-17 09:34:49.706117544 +0000 UTC m=+1511.806921043" observedRunningTime="2026-03-17 09:34:50.300590742 +0000 UTC m=+1512.401394251" watchObservedRunningTime="2026-03-17 09:34:50.300659414 +0000 UTC m=+1512.401462933" Mar 17 09:34:50 crc kubenswrapper[4813]: I0317 09:34:50.362879 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67b789f86c-r46bt" podStartSLOduration=3.3628635559999998 podStartE2EDuration="3.362863556s" podCreationTimestamp="2026-03-17 09:34:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:34:50.356889088 +0000 UTC m=+1512.457692587" watchObservedRunningTime="2026-03-17 09:34:50.362863556 +0000 UTC m=+1512.463667055" Mar 17 09:34:51 crc kubenswrapper[4813]: I0317 09:34:51.301760 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5vxx" event={"ID":"ee189143-6f3d-489d-be41-604ee0f72772","Type":"ContainerStarted","Data":"a094f443d8a94e25518a83a80bb74edd72ecdeeab21b4025f99f0ebcb5846b70"} Mar 17 09:34:51 crc kubenswrapper[4813]: I0317 09:34:51.334719 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-w5vxx" podStartSLOduration=2.763845724 podStartE2EDuration="10.334701821s" podCreationTimestamp="2026-03-17 09:34:41 +0000 UTC" firstStartedPulling="2026-03-17 09:34:43.156059909 +0000 UTC m=+1505.256863408" lastFinishedPulling="2026-03-17 09:34:50.726916006 +0000 UTC m=+1512.827719505" observedRunningTime="2026-03-17 09:34:51.329855888 +0000 UTC m=+1513.430659427" watchObservedRunningTime="2026-03-17 09:34:51.334701821 +0000 UTC m=+1513.435505320" Mar 17 09:34:51 crc kubenswrapper[4813]: I0317 09:34:51.666222 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-w5vxx" Mar 17 09:34:51 crc kubenswrapper[4813]: I0317 09:34:51.666299 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-w5vxx" Mar 17 09:34:52 crc kubenswrapper[4813]: I0317 09:34:52.714388 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-w5vxx" podUID="ee189143-6f3d-489d-be41-604ee0f72772" containerName="registry-server" probeResult="failure" output=< Mar 17 09:34:52 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Mar 17 09:34:52 crc kubenswrapper[4813]: > Mar 17 09:34:55 crc kubenswrapper[4813]: I0317 09:34:55.808902 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-8pbz7" Mar 17 09:34:55 crc kubenswrapper[4813]: I0317 09:34:55.809313 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-8pbz7" Mar 17 09:34:55 crc kubenswrapper[4813]: I0317 09:34:55.869556 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-8pbz7" Mar 17 09:34:56 crc kubenswrapper[4813]: I0317 09:34:56.426779 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-8pbz7" Mar 17 09:34:56 crc kubenswrapper[4813]: I0317 09:34:56.488447 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8pbz7"] Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.399180 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-8pbz7" podUID="1116d60a-9394-4311-ab91-fa9181c8dd66" containerName="registry-server" containerID="cri-o://6329b5e6bfca5977e464952a606bf8051ce46982f04a378ac87acff18c020c9a" gracePeriod=2 Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.399803 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.461532 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-cv6cc"] Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.462049 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" podUID="0b1a228b-5bd8-4340-a286-cf609851c75d" containerName="dnsmasq-dns" containerID="cri-o://c1a329c4c5592bc62a8a342e76504c2ee24fde54a809235d3f5d2308031e86bf" gracePeriod=10 Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.631522 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-drzd4"] Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.633176 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.655130 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cceef96-6dbd-4173-b363-3bc4fbf4f598-config\") pod \"dnsmasq-dns-cb6ffcf87-drzd4\" (UID: \"3cceef96-6dbd-4173-b363-3bc4fbf4f598\") " pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.655184 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3cceef96-6dbd-4173-b363-3bc4fbf4f598-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-drzd4\" (UID: \"3cceef96-6dbd-4173-b363-3bc4fbf4f598\") " pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.655267 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3cceef96-6dbd-4173-b363-3bc4fbf4f598-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-drzd4\" (UID: \"3cceef96-6dbd-4173-b363-3bc4fbf4f598\") " pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.655328 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3cceef96-6dbd-4173-b363-3bc4fbf4f598-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-drzd4\" (UID: \"3cceef96-6dbd-4173-b363-3bc4fbf4f598\") " pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.656166 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3cceef96-6dbd-4173-b363-3bc4fbf4f598-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-drzd4\" (UID: \"3cceef96-6dbd-4173-b363-3bc4fbf4f598\") " pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.656360 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhlz5\" (UniqueName: \"kubernetes.io/projected/3cceef96-6dbd-4173-b363-3bc4fbf4f598-kube-api-access-jhlz5\") pod \"dnsmasq-dns-cb6ffcf87-drzd4\" (UID: \"3cceef96-6dbd-4173-b363-3bc4fbf4f598\") " pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.656437 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3cceef96-6dbd-4173-b363-3bc4fbf4f598-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-drzd4\" (UID: \"3cceef96-6dbd-4173-b363-3bc4fbf4f598\") " pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.663512 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-drzd4"] Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.761188 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhlz5\" (UniqueName: \"kubernetes.io/projected/3cceef96-6dbd-4173-b363-3bc4fbf4f598-kube-api-access-jhlz5\") pod \"dnsmasq-dns-cb6ffcf87-drzd4\" (UID: \"3cceef96-6dbd-4173-b363-3bc4fbf4f598\") " pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.761259 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3cceef96-6dbd-4173-b363-3bc4fbf4f598-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-drzd4\" (UID: \"3cceef96-6dbd-4173-b363-3bc4fbf4f598\") " pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.761338 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cceef96-6dbd-4173-b363-3bc4fbf4f598-config\") pod \"dnsmasq-dns-cb6ffcf87-drzd4\" (UID: \"3cceef96-6dbd-4173-b363-3bc4fbf4f598\") " pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.761373 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3cceef96-6dbd-4173-b363-3bc4fbf4f598-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-drzd4\" (UID: \"3cceef96-6dbd-4173-b363-3bc4fbf4f598\") " pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.761403 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3cceef96-6dbd-4173-b363-3bc4fbf4f598-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-drzd4\" (UID: \"3cceef96-6dbd-4173-b363-3bc4fbf4f598\") " pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.761431 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3cceef96-6dbd-4173-b363-3bc4fbf4f598-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-drzd4\" (UID: \"3cceef96-6dbd-4173-b363-3bc4fbf4f598\") " pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.761485 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3cceef96-6dbd-4173-b363-3bc4fbf4f598-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-drzd4\" (UID: \"3cceef96-6dbd-4173-b363-3bc4fbf4f598\") " pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.762418 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3cceef96-6dbd-4173-b363-3bc4fbf4f598-ovsdbserver-sb\") pod \"dnsmasq-dns-cb6ffcf87-drzd4\" (UID: \"3cceef96-6dbd-4173-b363-3bc4fbf4f598\") " pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.763320 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3cceef96-6dbd-4173-b363-3bc4fbf4f598-dns-swift-storage-0\") pod \"dnsmasq-dns-cb6ffcf87-drzd4\" (UID: \"3cceef96-6dbd-4173-b363-3bc4fbf4f598\") " pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.763404 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3cceef96-6dbd-4173-b363-3bc4fbf4f598-dns-svc\") pod \"dnsmasq-dns-cb6ffcf87-drzd4\" (UID: \"3cceef96-6dbd-4173-b363-3bc4fbf4f598\") " pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.763434 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3cceef96-6dbd-4173-b363-3bc4fbf4f598-ovsdbserver-nb\") pod \"dnsmasq-dns-cb6ffcf87-drzd4\" (UID: \"3cceef96-6dbd-4173-b363-3bc4fbf4f598\") " pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.763904 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/3cceef96-6dbd-4173-b363-3bc4fbf4f598-openstack-edpm-ipam\") pod \"dnsmasq-dns-cb6ffcf87-drzd4\" (UID: \"3cceef96-6dbd-4173-b363-3bc4fbf4f598\") " pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.764026 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3cceef96-6dbd-4173-b363-3bc4fbf4f598-config\") pod \"dnsmasq-dns-cb6ffcf87-drzd4\" (UID: \"3cceef96-6dbd-4173-b363-3bc4fbf4f598\") " pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.789774 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhlz5\" (UniqueName: \"kubernetes.io/projected/3cceef96-6dbd-4173-b363-3bc4fbf4f598-kube-api-access-jhlz5\") pod \"dnsmasq-dns-cb6ffcf87-drzd4\" (UID: \"3cceef96-6dbd-4173-b363-3bc4fbf4f598\") " pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:34:59 crc kubenswrapper[4813]: I0317 09:34:59.962390 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.035469 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8pbz7" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.052673 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.065735 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1116d60a-9394-4311-ab91-fa9181c8dd66-catalog-content\") pod \"1116d60a-9394-4311-ab91-fa9181c8dd66\" (UID: \"1116d60a-9394-4311-ab91-fa9181c8dd66\") " Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.065803 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1116d60a-9394-4311-ab91-fa9181c8dd66-utilities\") pod \"1116d60a-9394-4311-ab91-fa9181c8dd66\" (UID: \"1116d60a-9394-4311-ab91-fa9181c8dd66\") " Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.065880 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mn8sp\" (UniqueName: \"kubernetes.io/projected/1116d60a-9394-4311-ab91-fa9181c8dd66-kube-api-access-mn8sp\") pod \"1116d60a-9394-4311-ab91-fa9181c8dd66\" (UID: \"1116d60a-9394-4311-ab91-fa9181c8dd66\") " Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.066891 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1116d60a-9394-4311-ab91-fa9181c8dd66-utilities" (OuterVolumeSpecName: "utilities") pod "1116d60a-9394-4311-ab91-fa9181c8dd66" (UID: "1116d60a-9394-4311-ab91-fa9181c8dd66"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.073170 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1116d60a-9394-4311-ab91-fa9181c8dd66-kube-api-access-mn8sp" (OuterVolumeSpecName: "kube-api-access-mn8sp") pod "1116d60a-9394-4311-ab91-fa9181c8dd66" (UID: "1116d60a-9394-4311-ab91-fa9181c8dd66"). InnerVolumeSpecName "kube-api-access-mn8sp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.169216 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cv6nf\" (UniqueName: \"kubernetes.io/projected/0b1a228b-5bd8-4340-a286-cf609851c75d-kube-api-access-cv6nf\") pod \"0b1a228b-5bd8-4340-a286-cf609851c75d\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.169590 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-ovsdbserver-nb\") pod \"0b1a228b-5bd8-4340-a286-cf609851c75d\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.169639 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-dns-svc\") pod \"0b1a228b-5bd8-4340-a286-cf609851c75d\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.169679 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-config\") pod \"0b1a228b-5bd8-4340-a286-cf609851c75d\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.169705 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-ovsdbserver-sb\") pod \"0b1a228b-5bd8-4340-a286-cf609851c75d\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.169741 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-dns-swift-storage-0\") pod \"0b1a228b-5bd8-4340-a286-cf609851c75d\" (UID: \"0b1a228b-5bd8-4340-a286-cf609851c75d\") " Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.170211 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1116d60a-9394-4311-ab91-fa9181c8dd66-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.170224 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mn8sp\" (UniqueName: \"kubernetes.io/projected/1116d60a-9394-4311-ab91-fa9181c8dd66-kube-api-access-mn8sp\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.170917 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1116d60a-9394-4311-ab91-fa9181c8dd66-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1116d60a-9394-4311-ab91-fa9181c8dd66" (UID: "1116d60a-9394-4311-ab91-fa9181c8dd66"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.176843 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b1a228b-5bd8-4340-a286-cf609851c75d-kube-api-access-cv6nf" (OuterVolumeSpecName: "kube-api-access-cv6nf") pod "0b1a228b-5bd8-4340-a286-cf609851c75d" (UID: "0b1a228b-5bd8-4340-a286-cf609851c75d"). InnerVolumeSpecName "kube-api-access-cv6nf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.219269 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "0b1a228b-5bd8-4340-a286-cf609851c75d" (UID: "0b1a228b-5bd8-4340-a286-cf609851c75d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.222070 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "0b1a228b-5bd8-4340-a286-cf609851c75d" (UID: "0b1a228b-5bd8-4340-a286-cf609851c75d"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.224481 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "0b1a228b-5bd8-4340-a286-cf609851c75d" (UID: "0b1a228b-5bd8-4340-a286-cf609851c75d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.241997 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "0b1a228b-5bd8-4340-a286-cf609851c75d" (UID: "0b1a228b-5bd8-4340-a286-cf609851c75d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.249572 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cb6ffcf87-drzd4"] Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.250866 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-config" (OuterVolumeSpecName: "config") pod "0b1a228b-5bd8-4340-a286-cf609851c75d" (UID: "0b1a228b-5bd8-4340-a286-cf609851c75d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.273376 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1116d60a-9394-4311-ab91-fa9181c8dd66-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.273428 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cv6nf\" (UniqueName: \"kubernetes.io/projected/0b1a228b-5bd8-4340-a286-cf609851c75d-kube-api-access-cv6nf\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.273442 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.273451 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.273460 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.273672 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.273683 4813 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/0b1a228b-5bd8-4340-a286-cf609851c75d-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.373278 4813 generic.go:334] "Generic (PLEG): container finished" podID="0b1a228b-5bd8-4340-a286-cf609851c75d" containerID="c1a329c4c5592bc62a8a342e76504c2ee24fde54a809235d3f5d2308031e86bf" exitCode=0 Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.373339 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" event={"ID":"0b1a228b-5bd8-4340-a286-cf609851c75d","Type":"ContainerDied","Data":"c1a329c4c5592bc62a8a342e76504c2ee24fde54a809235d3f5d2308031e86bf"} Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.373364 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" event={"ID":"0b1a228b-5bd8-4340-a286-cf609851c75d","Type":"ContainerDied","Data":"e2570020dc39abf30bcb2883a46a66b3ee8e5e4dece7b1df0d986e54f24b5437"} Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.373381 4813 scope.go:117] "RemoveContainer" containerID="c1a329c4c5592bc62a8a342e76504c2ee24fde54a809235d3f5d2308031e86bf" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.373394 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-59cf4bdb65-cv6cc" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.378257 4813 generic.go:334] "Generic (PLEG): container finished" podID="1116d60a-9394-4311-ab91-fa9181c8dd66" containerID="6329b5e6bfca5977e464952a606bf8051ce46982f04a378ac87acff18c020c9a" exitCode=0 Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.378318 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8pbz7" event={"ID":"1116d60a-9394-4311-ab91-fa9181c8dd66","Type":"ContainerDied","Data":"6329b5e6bfca5977e464952a606bf8051ce46982f04a378ac87acff18c020c9a"} Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.378334 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-8pbz7" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.378339 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-8pbz7" event={"ID":"1116d60a-9394-4311-ab91-fa9181c8dd66","Type":"ContainerDied","Data":"32f4aa866b8d9fbe75e9df5419efc5ba185675147f808adb6ea4cc087e2fee72"} Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.379588 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" event={"ID":"3cceef96-6dbd-4173-b363-3bc4fbf4f598","Type":"ContainerStarted","Data":"cd094729ba7c8c39e050c525ce29d6e2be424b380e513f91f0443612beed6f08"} Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.409258 4813 scope.go:117] "RemoveContainer" containerID="bc6ab5b77103f301ace06b4bf22d90de46dbfb3bd748a19e1774b5d8eceb994d" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.428656 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-cv6cc"] Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.440567 4813 scope.go:117] "RemoveContainer" containerID="c1a329c4c5592bc62a8a342e76504c2ee24fde54a809235d3f5d2308031e86bf" Mar 17 09:35:00 crc kubenswrapper[4813]: E0317 09:35:00.441028 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1a329c4c5592bc62a8a342e76504c2ee24fde54a809235d3f5d2308031e86bf\": container with ID starting with c1a329c4c5592bc62a8a342e76504c2ee24fde54a809235d3f5d2308031e86bf not found: ID does not exist" containerID="c1a329c4c5592bc62a8a342e76504c2ee24fde54a809235d3f5d2308031e86bf" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.441095 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1a329c4c5592bc62a8a342e76504c2ee24fde54a809235d3f5d2308031e86bf"} err="failed to get container status \"c1a329c4c5592bc62a8a342e76504c2ee24fde54a809235d3f5d2308031e86bf\": rpc error: code = NotFound desc = could not find container \"c1a329c4c5592bc62a8a342e76504c2ee24fde54a809235d3f5d2308031e86bf\": container with ID starting with c1a329c4c5592bc62a8a342e76504c2ee24fde54a809235d3f5d2308031e86bf not found: ID does not exist" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.441128 4813 scope.go:117] "RemoveContainer" containerID="bc6ab5b77103f301ace06b4bf22d90de46dbfb3bd748a19e1774b5d8eceb994d" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.441258 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-59cf4bdb65-cv6cc"] Mar 17 09:35:00 crc kubenswrapper[4813]: E0317 09:35:00.441581 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc6ab5b77103f301ace06b4bf22d90de46dbfb3bd748a19e1774b5d8eceb994d\": container with ID starting with bc6ab5b77103f301ace06b4bf22d90de46dbfb3bd748a19e1774b5d8eceb994d not found: ID does not exist" containerID="bc6ab5b77103f301ace06b4bf22d90de46dbfb3bd748a19e1774b5d8eceb994d" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.441661 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc6ab5b77103f301ace06b4bf22d90de46dbfb3bd748a19e1774b5d8eceb994d"} err="failed to get container status \"bc6ab5b77103f301ace06b4bf22d90de46dbfb3bd748a19e1774b5d8eceb994d\": rpc error: code = NotFound desc = could not find container \"bc6ab5b77103f301ace06b4bf22d90de46dbfb3bd748a19e1774b5d8eceb994d\": container with ID starting with bc6ab5b77103f301ace06b4bf22d90de46dbfb3bd748a19e1774b5d8eceb994d not found: ID does not exist" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.441703 4813 scope.go:117] "RemoveContainer" containerID="6329b5e6bfca5977e464952a606bf8051ce46982f04a378ac87acff18c020c9a" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.453793 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-8pbz7"] Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.462421 4813 scope.go:117] "RemoveContainer" containerID="a5b930d6471e1aa9daf3e525cc453d1d99d6129b90d6d09f93265174de9ddc2a" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.464162 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-8pbz7"] Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.481468 4813 scope.go:117] "RemoveContainer" containerID="5cff32a263fac1ab39538939589f2a42dd0a96beaadfb8e9e90b22218998819b" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.500761 4813 scope.go:117] "RemoveContainer" containerID="6329b5e6bfca5977e464952a606bf8051ce46982f04a378ac87acff18c020c9a" Mar 17 09:35:00 crc kubenswrapper[4813]: E0317 09:35:00.501208 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6329b5e6bfca5977e464952a606bf8051ce46982f04a378ac87acff18c020c9a\": container with ID starting with 6329b5e6bfca5977e464952a606bf8051ce46982f04a378ac87acff18c020c9a not found: ID does not exist" containerID="6329b5e6bfca5977e464952a606bf8051ce46982f04a378ac87acff18c020c9a" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.501253 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6329b5e6bfca5977e464952a606bf8051ce46982f04a378ac87acff18c020c9a"} err="failed to get container status \"6329b5e6bfca5977e464952a606bf8051ce46982f04a378ac87acff18c020c9a\": rpc error: code = NotFound desc = could not find container \"6329b5e6bfca5977e464952a606bf8051ce46982f04a378ac87acff18c020c9a\": container with ID starting with 6329b5e6bfca5977e464952a606bf8051ce46982f04a378ac87acff18c020c9a not found: ID does not exist" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.501303 4813 scope.go:117] "RemoveContainer" containerID="a5b930d6471e1aa9daf3e525cc453d1d99d6129b90d6d09f93265174de9ddc2a" Mar 17 09:35:00 crc kubenswrapper[4813]: E0317 09:35:00.502180 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a5b930d6471e1aa9daf3e525cc453d1d99d6129b90d6d09f93265174de9ddc2a\": container with ID starting with a5b930d6471e1aa9daf3e525cc453d1d99d6129b90d6d09f93265174de9ddc2a not found: ID does not exist" containerID="a5b930d6471e1aa9daf3e525cc453d1d99d6129b90d6d09f93265174de9ddc2a" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.502216 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a5b930d6471e1aa9daf3e525cc453d1d99d6129b90d6d09f93265174de9ddc2a"} err="failed to get container status \"a5b930d6471e1aa9daf3e525cc453d1d99d6129b90d6d09f93265174de9ddc2a\": rpc error: code = NotFound desc = could not find container \"a5b930d6471e1aa9daf3e525cc453d1d99d6129b90d6d09f93265174de9ddc2a\": container with ID starting with a5b930d6471e1aa9daf3e525cc453d1d99d6129b90d6d09f93265174de9ddc2a not found: ID does not exist" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.502245 4813 scope.go:117] "RemoveContainer" containerID="5cff32a263fac1ab39538939589f2a42dd0a96beaadfb8e9e90b22218998819b" Mar 17 09:35:00 crc kubenswrapper[4813]: E0317 09:35:00.504003 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cff32a263fac1ab39538939589f2a42dd0a96beaadfb8e9e90b22218998819b\": container with ID starting with 5cff32a263fac1ab39538939589f2a42dd0a96beaadfb8e9e90b22218998819b not found: ID does not exist" containerID="5cff32a263fac1ab39538939589f2a42dd0a96beaadfb8e9e90b22218998819b" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.504034 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cff32a263fac1ab39538939589f2a42dd0a96beaadfb8e9e90b22218998819b"} err="failed to get container status \"5cff32a263fac1ab39538939589f2a42dd0a96beaadfb8e9e90b22218998819b\": rpc error: code = NotFound desc = could not find container \"5cff32a263fac1ab39538939589f2a42dd0a96beaadfb8e9e90b22218998819b\": container with ID starting with 5cff32a263fac1ab39538939589f2a42dd0a96beaadfb8e9e90b22218998819b not found: ID does not exist" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.744753 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b1a228b-5bd8-4340-a286-cf609851c75d" path="/var/lib/kubelet/pods/0b1a228b-5bd8-4340-a286-cf609851c75d/volumes" Mar 17 09:35:00 crc kubenswrapper[4813]: I0317 09:35:00.745631 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1116d60a-9394-4311-ab91-fa9181c8dd66" path="/var/lib/kubelet/pods/1116d60a-9394-4311-ab91-fa9181c8dd66/volumes" Mar 17 09:35:01 crc kubenswrapper[4813]: I0317 09:35:01.393574 4813 generic.go:334] "Generic (PLEG): container finished" podID="3cceef96-6dbd-4173-b363-3bc4fbf4f598" containerID="524053a76ebf2e4e407d0947efe7e2bc2c87c016e6384c11724c578a095fcb5c" exitCode=0 Mar 17 09:35:01 crc kubenswrapper[4813]: I0317 09:35:01.393634 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" event={"ID":"3cceef96-6dbd-4173-b363-3bc4fbf4f598","Type":"ContainerDied","Data":"524053a76ebf2e4e407d0947efe7e2bc2c87c016e6384c11724c578a095fcb5c"} Mar 17 09:35:01 crc kubenswrapper[4813]: I0317 09:35:01.709755 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-w5vxx" Mar 17 09:35:01 crc kubenswrapper[4813]: I0317 09:35:01.782848 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-w5vxx" Mar 17 09:35:02 crc kubenswrapper[4813]: I0317 09:35:02.270460 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w5vxx"] Mar 17 09:35:02 crc kubenswrapper[4813]: I0317 09:35:02.407435 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" event={"ID":"3cceef96-6dbd-4173-b363-3bc4fbf4f598","Type":"ContainerStarted","Data":"b9f31d885b6886cd16c1275e91a7699a7d8b93f3cc8f85c0846da72e51a305ba"} Mar 17 09:35:02 crc kubenswrapper[4813]: I0317 09:35:02.437672 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" podStartSLOduration=3.437651734 podStartE2EDuration="3.437651734s" podCreationTimestamp="2026-03-17 09:34:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:35:02.431913214 +0000 UTC m=+1524.532716743" watchObservedRunningTime="2026-03-17 09:35:02.437651734 +0000 UTC m=+1524.538455233" Mar 17 09:35:03 crc kubenswrapper[4813]: I0317 09:35:03.433880 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:35:03 crc kubenswrapper[4813]: I0317 09:35:03.434413 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-w5vxx" podUID="ee189143-6f3d-489d-be41-604ee0f72772" containerName="registry-server" containerID="cri-o://a094f443d8a94e25518a83a80bb74edd72ecdeeab21b4025f99f0ebcb5846b70" gracePeriod=2 Mar 17 09:35:03 crc kubenswrapper[4813]: I0317 09:35:03.977678 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w5vxx" Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.082186 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee189143-6f3d-489d-be41-604ee0f72772-catalog-content\") pod \"ee189143-6f3d-489d-be41-604ee0f72772\" (UID: \"ee189143-6f3d-489d-be41-604ee0f72772\") " Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.082342 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee189143-6f3d-489d-be41-604ee0f72772-utilities\") pod \"ee189143-6f3d-489d-be41-604ee0f72772\" (UID: \"ee189143-6f3d-489d-be41-604ee0f72772\") " Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.082463 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dkht2\" (UniqueName: \"kubernetes.io/projected/ee189143-6f3d-489d-be41-604ee0f72772-kube-api-access-dkht2\") pod \"ee189143-6f3d-489d-be41-604ee0f72772\" (UID: \"ee189143-6f3d-489d-be41-604ee0f72772\") " Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.083377 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee189143-6f3d-489d-be41-604ee0f72772-utilities" (OuterVolumeSpecName: "utilities") pod "ee189143-6f3d-489d-be41-604ee0f72772" (UID: "ee189143-6f3d-489d-be41-604ee0f72772"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.088999 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee189143-6f3d-489d-be41-604ee0f72772-kube-api-access-dkht2" (OuterVolumeSpecName: "kube-api-access-dkht2") pod "ee189143-6f3d-489d-be41-604ee0f72772" (UID: "ee189143-6f3d-489d-be41-604ee0f72772"). InnerVolumeSpecName "kube-api-access-dkht2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.184994 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee189143-6f3d-489d-be41-604ee0f72772-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.185031 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dkht2\" (UniqueName: \"kubernetes.io/projected/ee189143-6f3d-489d-be41-604ee0f72772-kube-api-access-dkht2\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.248518 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee189143-6f3d-489d-be41-604ee0f72772-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee189143-6f3d-489d-be41-604ee0f72772" (UID: "ee189143-6f3d-489d-be41-604ee0f72772"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.287506 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee189143-6f3d-489d-be41-604ee0f72772-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.449322 4813 generic.go:334] "Generic (PLEG): container finished" podID="ee189143-6f3d-489d-be41-604ee0f72772" containerID="a094f443d8a94e25518a83a80bb74edd72ecdeeab21b4025f99f0ebcb5846b70" exitCode=0 Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.449418 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-w5vxx" Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.449435 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5vxx" event={"ID":"ee189143-6f3d-489d-be41-604ee0f72772","Type":"ContainerDied","Data":"a094f443d8a94e25518a83a80bb74edd72ecdeeab21b4025f99f0ebcb5846b70"} Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.449510 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-w5vxx" event={"ID":"ee189143-6f3d-489d-be41-604ee0f72772","Type":"ContainerDied","Data":"92deef0c321d544683d9b8e7d8cf3452850f0826efa913fc47004b16e5d62156"} Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.449539 4813 scope.go:117] "RemoveContainer" containerID="a094f443d8a94e25518a83a80bb74edd72ecdeeab21b4025f99f0ebcb5846b70" Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.505350 4813 scope.go:117] "RemoveContainer" containerID="a711b4fe15101b0fc9a50d6790cb354c4d562f0c35a172cee29b51958a8195a6" Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.514094 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-w5vxx"] Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.525499 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-w5vxx"] Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.549033 4813 scope.go:117] "RemoveContainer" containerID="ed7212d277e75ed05c93691ac8c9504193a643c1d280b36a30e51ce6a0b7e14a" Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.606952 4813 scope.go:117] "RemoveContainer" containerID="a094f443d8a94e25518a83a80bb74edd72ecdeeab21b4025f99f0ebcb5846b70" Mar 17 09:35:04 crc kubenswrapper[4813]: E0317 09:35:04.607442 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a094f443d8a94e25518a83a80bb74edd72ecdeeab21b4025f99f0ebcb5846b70\": container with ID starting with a094f443d8a94e25518a83a80bb74edd72ecdeeab21b4025f99f0ebcb5846b70 not found: ID does not exist" containerID="a094f443d8a94e25518a83a80bb74edd72ecdeeab21b4025f99f0ebcb5846b70" Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.607492 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a094f443d8a94e25518a83a80bb74edd72ecdeeab21b4025f99f0ebcb5846b70"} err="failed to get container status \"a094f443d8a94e25518a83a80bb74edd72ecdeeab21b4025f99f0ebcb5846b70\": rpc error: code = NotFound desc = could not find container \"a094f443d8a94e25518a83a80bb74edd72ecdeeab21b4025f99f0ebcb5846b70\": container with ID starting with a094f443d8a94e25518a83a80bb74edd72ecdeeab21b4025f99f0ebcb5846b70 not found: ID does not exist" Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.607526 4813 scope.go:117] "RemoveContainer" containerID="a711b4fe15101b0fc9a50d6790cb354c4d562f0c35a172cee29b51958a8195a6" Mar 17 09:35:04 crc kubenswrapper[4813]: E0317 09:35:04.608048 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a711b4fe15101b0fc9a50d6790cb354c4d562f0c35a172cee29b51958a8195a6\": container with ID starting with a711b4fe15101b0fc9a50d6790cb354c4d562f0c35a172cee29b51958a8195a6 not found: ID does not exist" containerID="a711b4fe15101b0fc9a50d6790cb354c4d562f0c35a172cee29b51958a8195a6" Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.608073 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a711b4fe15101b0fc9a50d6790cb354c4d562f0c35a172cee29b51958a8195a6"} err="failed to get container status \"a711b4fe15101b0fc9a50d6790cb354c4d562f0c35a172cee29b51958a8195a6\": rpc error: code = NotFound desc = could not find container \"a711b4fe15101b0fc9a50d6790cb354c4d562f0c35a172cee29b51958a8195a6\": container with ID starting with a711b4fe15101b0fc9a50d6790cb354c4d562f0c35a172cee29b51958a8195a6 not found: ID does not exist" Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.608089 4813 scope.go:117] "RemoveContainer" containerID="ed7212d277e75ed05c93691ac8c9504193a643c1d280b36a30e51ce6a0b7e14a" Mar 17 09:35:04 crc kubenswrapper[4813]: E0317 09:35:04.608504 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed7212d277e75ed05c93691ac8c9504193a643c1d280b36a30e51ce6a0b7e14a\": container with ID starting with ed7212d277e75ed05c93691ac8c9504193a643c1d280b36a30e51ce6a0b7e14a not found: ID does not exist" containerID="ed7212d277e75ed05c93691ac8c9504193a643c1d280b36a30e51ce6a0b7e14a" Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.608529 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed7212d277e75ed05c93691ac8c9504193a643c1d280b36a30e51ce6a0b7e14a"} err="failed to get container status \"ed7212d277e75ed05c93691ac8c9504193a643c1d280b36a30e51ce6a0b7e14a\": rpc error: code = NotFound desc = could not find container \"ed7212d277e75ed05c93691ac8c9504193a643c1d280b36a30e51ce6a0b7e14a\": container with ID starting with ed7212d277e75ed05c93691ac8c9504193a643c1d280b36a30e51ce6a0b7e14a not found: ID does not exist" Mar 17 09:35:04 crc kubenswrapper[4813]: I0317 09:35:04.743578 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee189143-6f3d-489d-be41-604ee0f72772" path="/var/lib/kubelet/pods/ee189143-6f3d-489d-be41-604ee0f72772/volumes" Mar 17 09:35:09 crc kubenswrapper[4813]: I0317 09:35:09.965890 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cb6ffcf87-drzd4" Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.069624 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-r46bt"] Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.070355 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-67b789f86c-r46bt" podUID="af669e24-da1e-4585-865c-e6d978c12838" containerName="dnsmasq-dns" containerID="cri-o://f95998cbf5bb06f0793dbfffb2b455c0da8571e30177362a52d817b3aefc91d8" gracePeriod=10 Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.526983 4813 generic.go:334] "Generic (PLEG): container finished" podID="af669e24-da1e-4585-865c-e6d978c12838" containerID="f95998cbf5bb06f0793dbfffb2b455c0da8571e30177362a52d817b3aefc91d8" exitCode=0 Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.527133 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-r46bt" event={"ID":"af669e24-da1e-4585-865c-e6d978c12838","Type":"ContainerDied","Data":"f95998cbf5bb06f0793dbfffb2b455c0da8571e30177362a52d817b3aefc91d8"} Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.527350 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67b789f86c-r46bt" event={"ID":"af669e24-da1e-4585-865c-e6d978c12838","Type":"ContainerDied","Data":"635abe893adfd1281af877443424901a91f190c1edac8822b91241e09224c847"} Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.527366 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="635abe893adfd1281af877443424901a91f190c1edac8822b91241e09224c847" Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.612424 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.737222 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-config\") pod \"af669e24-da1e-4585-865c-e6d978c12838\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.737413 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5f5mz\" (UniqueName: \"kubernetes.io/projected/af669e24-da1e-4585-865c-e6d978c12838-kube-api-access-5f5mz\") pod \"af669e24-da1e-4585-865c-e6d978c12838\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.737477 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-ovsdbserver-nb\") pod \"af669e24-da1e-4585-865c-e6d978c12838\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.737572 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-dns-svc\") pod \"af669e24-da1e-4585-865c-e6d978c12838\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.738097 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-openstack-edpm-ipam\") pod \"af669e24-da1e-4585-865c-e6d978c12838\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.738173 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-dns-swift-storage-0\") pod \"af669e24-da1e-4585-865c-e6d978c12838\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.738205 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-ovsdbserver-sb\") pod \"af669e24-da1e-4585-865c-e6d978c12838\" (UID: \"af669e24-da1e-4585-865c-e6d978c12838\") " Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.743078 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af669e24-da1e-4585-865c-e6d978c12838-kube-api-access-5f5mz" (OuterVolumeSpecName: "kube-api-access-5f5mz") pod "af669e24-da1e-4585-865c-e6d978c12838" (UID: "af669e24-da1e-4585-865c-e6d978c12838"). InnerVolumeSpecName "kube-api-access-5f5mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.787938 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "af669e24-da1e-4585-865c-e6d978c12838" (UID: "af669e24-da1e-4585-865c-e6d978c12838"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.797161 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "af669e24-da1e-4585-865c-e6d978c12838" (UID: "af669e24-da1e-4585-865c-e6d978c12838"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.802337 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "af669e24-da1e-4585-865c-e6d978c12838" (UID: "af669e24-da1e-4585-865c-e6d978c12838"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.803984 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "af669e24-da1e-4585-865c-e6d978c12838" (UID: "af669e24-da1e-4585-865c-e6d978c12838"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.818255 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-config" (OuterVolumeSpecName: "config") pod "af669e24-da1e-4585-865c-e6d978c12838" (UID: "af669e24-da1e-4585-865c-e6d978c12838"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.822789 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "af669e24-da1e-4585-865c-e6d978c12838" (UID: "af669e24-da1e-4585-865c-e6d978c12838"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.840780 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5f5mz\" (UniqueName: \"kubernetes.io/projected/af669e24-da1e-4585-865c-e6d978c12838-kube-api-access-5f5mz\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.840817 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.840847 4813 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-dns-svc\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.840859 4813 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.840871 4813 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.840882 4813 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:10 crc kubenswrapper[4813]: I0317 09:35:10.840894 4813 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af669e24-da1e-4585-865c-e6d978c12838-config\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:11 crc kubenswrapper[4813]: I0317 09:35:11.542648 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67b789f86c-r46bt" Mar 17 09:35:11 crc kubenswrapper[4813]: I0317 09:35:11.593215 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-r46bt"] Mar 17 09:35:11 crc kubenswrapper[4813]: I0317 09:35:11.609471 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-67b789f86c-r46bt"] Mar 17 09:35:12 crc kubenswrapper[4813]: I0317 09:35:12.747214 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af669e24-da1e-4585-865c-e6d978c12838" path="/var/lib/kubelet/pods/af669e24-da1e-4585-865c-e6d978c12838/volumes" Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.661423 4813 generic.go:334] "Generic (PLEG): container finished" podID="16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b" containerID="a1e83709dd2273fea789e5921d2b30f17f4925695ba5ce7811df41ecec5bd30c" exitCode=0 Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.661544 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b","Type":"ContainerDied","Data":"a1e83709dd2273fea789e5921d2b30f17f4925695ba5ce7811df41ecec5bd30c"} Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.967740 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr"] Mar 17 09:35:21 crc kubenswrapper[4813]: E0317 09:35:21.969772 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1116d60a-9394-4311-ab91-fa9181c8dd66" containerName="extract-utilities" Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.969802 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1116d60a-9394-4311-ab91-fa9181c8dd66" containerName="extract-utilities" Mar 17 09:35:21 crc kubenswrapper[4813]: E0317 09:35:21.969840 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1116d60a-9394-4311-ab91-fa9181c8dd66" containerName="registry-server" Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.969849 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1116d60a-9394-4311-ab91-fa9181c8dd66" containerName="registry-server" Mar 17 09:35:21 crc kubenswrapper[4813]: E0317 09:35:21.969880 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af669e24-da1e-4585-865c-e6d978c12838" containerName="dnsmasq-dns" Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.969889 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="af669e24-da1e-4585-865c-e6d978c12838" containerName="dnsmasq-dns" Mar 17 09:35:21 crc kubenswrapper[4813]: E0317 09:35:21.969921 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af669e24-da1e-4585-865c-e6d978c12838" containerName="init" Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.969930 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="af669e24-da1e-4585-865c-e6d978c12838" containerName="init" Mar 17 09:35:21 crc kubenswrapper[4813]: E0317 09:35:21.969955 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee189143-6f3d-489d-be41-604ee0f72772" containerName="extract-utilities" Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.969962 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee189143-6f3d-489d-be41-604ee0f72772" containerName="extract-utilities" Mar 17 09:35:21 crc kubenswrapper[4813]: E0317 09:35:21.969974 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1116d60a-9394-4311-ab91-fa9181c8dd66" containerName="extract-content" Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.969983 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1116d60a-9394-4311-ab91-fa9181c8dd66" containerName="extract-content" Mar 17 09:35:21 crc kubenswrapper[4813]: E0317 09:35:21.970009 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee189143-6f3d-489d-be41-604ee0f72772" containerName="extract-content" Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.970017 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee189143-6f3d-489d-be41-604ee0f72772" containerName="extract-content" Mar 17 09:35:21 crc kubenswrapper[4813]: E0317 09:35:21.970056 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b1a228b-5bd8-4340-a286-cf609851c75d" containerName="dnsmasq-dns" Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.970065 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b1a228b-5bd8-4340-a286-cf609851c75d" containerName="dnsmasq-dns" Mar 17 09:35:21 crc kubenswrapper[4813]: E0317 09:35:21.970090 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee189143-6f3d-489d-be41-604ee0f72772" containerName="registry-server" Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.970098 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee189143-6f3d-489d-be41-604ee0f72772" containerName="registry-server" Mar 17 09:35:21 crc kubenswrapper[4813]: E0317 09:35:21.970112 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b1a228b-5bd8-4340-a286-cf609851c75d" containerName="init" Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.970121 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b1a228b-5bd8-4340-a286-cf609851c75d" containerName="init" Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.970760 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="af669e24-da1e-4585-865c-e6d978c12838" containerName="dnsmasq-dns" Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.970813 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="1116d60a-9394-4311-ab91-fa9181c8dd66" containerName="registry-server" Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.970845 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee189143-6f3d-489d-be41-604ee0f72772" containerName="registry-server" Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.970868 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b1a228b-5bd8-4340-a286-cf609851c75d" containerName="dnsmasq-dns" Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.971942 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr" Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.980141 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.980894 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hj9h4" Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.981369 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.981750 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 09:35:21 crc kubenswrapper[4813]: I0317 09:35:21.989058 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr"] Mar 17 09:35:22 crc kubenswrapper[4813]: I0317 09:35:22.077757 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v6n9v\" (UniqueName: \"kubernetes.io/projected/df99907b-2dfb-436f-b9a0-5c62dce82672-kube-api-access-v6n9v\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr\" (UID: \"df99907b-2dfb-436f-b9a0-5c62dce82672\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr" Mar 17 09:35:22 crc kubenswrapper[4813]: I0317 09:35:22.077874 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/df99907b-2dfb-436f-b9a0-5c62dce82672-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr\" (UID: \"df99907b-2dfb-436f-b9a0-5c62dce82672\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr" Mar 17 09:35:22 crc kubenswrapper[4813]: I0317 09:35:22.077919 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df99907b-2dfb-436f-b9a0-5c62dce82672-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr\" (UID: \"df99907b-2dfb-436f-b9a0-5c62dce82672\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr" Mar 17 09:35:22 crc kubenswrapper[4813]: I0317 09:35:22.078027 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df99907b-2dfb-436f-b9a0-5c62dce82672-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr\" (UID: \"df99907b-2dfb-436f-b9a0-5c62dce82672\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr" Mar 17 09:35:22 crc kubenswrapper[4813]: I0317 09:35:22.179097 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df99907b-2dfb-436f-b9a0-5c62dce82672-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr\" (UID: \"df99907b-2dfb-436f-b9a0-5c62dce82672\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr" Mar 17 09:35:22 crc kubenswrapper[4813]: I0317 09:35:22.179154 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v6n9v\" (UniqueName: \"kubernetes.io/projected/df99907b-2dfb-436f-b9a0-5c62dce82672-kube-api-access-v6n9v\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr\" (UID: \"df99907b-2dfb-436f-b9a0-5c62dce82672\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr" Mar 17 09:35:22 crc kubenswrapper[4813]: I0317 09:35:22.179209 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/df99907b-2dfb-436f-b9a0-5c62dce82672-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr\" (UID: \"df99907b-2dfb-436f-b9a0-5c62dce82672\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr" Mar 17 09:35:22 crc kubenswrapper[4813]: I0317 09:35:22.179251 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df99907b-2dfb-436f-b9a0-5c62dce82672-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr\" (UID: \"df99907b-2dfb-436f-b9a0-5c62dce82672\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr" Mar 17 09:35:22 crc kubenswrapper[4813]: I0317 09:35:22.183924 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df99907b-2dfb-436f-b9a0-5c62dce82672-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr\" (UID: \"df99907b-2dfb-436f-b9a0-5c62dce82672\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr" Mar 17 09:35:22 crc kubenswrapper[4813]: I0317 09:35:22.184004 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df99907b-2dfb-436f-b9a0-5c62dce82672-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr\" (UID: \"df99907b-2dfb-436f-b9a0-5c62dce82672\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr" Mar 17 09:35:22 crc kubenswrapper[4813]: I0317 09:35:22.184504 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/df99907b-2dfb-436f-b9a0-5c62dce82672-ssh-key-openstack-edpm-ipam\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr\" (UID: \"df99907b-2dfb-436f-b9a0-5c62dce82672\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr" Mar 17 09:35:22 crc kubenswrapper[4813]: I0317 09:35:22.207034 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v6n9v\" (UniqueName: \"kubernetes.io/projected/df99907b-2dfb-436f-b9a0-5c62dce82672-kube-api-access-v6n9v\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr\" (UID: \"df99907b-2dfb-436f-b9a0-5c62dce82672\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr" Mar 17 09:35:22 crc kubenswrapper[4813]: I0317 09:35:22.300211 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr" Mar 17 09:35:22 crc kubenswrapper[4813]: I0317 09:35:22.683388 4813 generic.go:334] "Generic (PLEG): container finished" podID="cfc379d0-d53a-4640-8d3f-a883db1aefa9" containerID="e6db8e4fb9e477f6fd0d51a9ab4a9b4a5f2770560a5526c186c89b4622d1afdd" exitCode=0 Mar 17 09:35:22 crc kubenswrapper[4813]: I0317 09:35:22.683811 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cfc379d0-d53a-4640-8d3f-a883db1aefa9","Type":"ContainerDied","Data":"e6db8e4fb9e477f6fd0d51a9ab4a9b4a5f2770560a5526c186c89b4622d1afdd"} Mar 17 09:35:22 crc kubenswrapper[4813]: I0317 09:35:22.689897 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b","Type":"ContainerStarted","Data":"81f1645c5cb7678e3618c1f8449fa953717f0d99025d8385bd10eb819f329ab4"} Mar 17 09:35:22 crc kubenswrapper[4813]: I0317 09:35:22.690172 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Mar 17 09:35:22 crc kubenswrapper[4813]: I0317 09:35:22.758233 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.758207709 podStartE2EDuration="36.758207709s" podCreationTimestamp="2026-03-17 09:34:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:35:22.74772333 +0000 UTC m=+1544.848526859" watchObservedRunningTime="2026-03-17 09:35:22.758207709 +0000 UTC m=+1544.859011218" Mar 17 09:35:22 crc kubenswrapper[4813]: I0317 09:35:22.891952 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr"] Mar 17 09:35:23 crc kubenswrapper[4813]: I0317 09:35:23.701408 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr" event={"ID":"df99907b-2dfb-436f-b9a0-5c62dce82672","Type":"ContainerStarted","Data":"e3dd7fb5c44ecf6db1c44dbb5603dd165da18925cb8007df3c16fdc1fa5da92e"} Mar 17 09:35:23 crc kubenswrapper[4813]: I0317 09:35:23.703958 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"cfc379d0-d53a-4640-8d3f-a883db1aefa9","Type":"ContainerStarted","Data":"9529fcc9831d00b146f65293b85822f8b0d06cbfead415c513618b3e128ada54"} Mar 17 09:35:23 crc kubenswrapper[4813]: I0317 09:35:23.735366 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.735351 podStartE2EDuration="36.735351s" podCreationTimestamp="2026-03-17 09:34:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 09:35:23.733490052 +0000 UTC m=+1545.834293601" watchObservedRunningTime="2026-03-17 09:35:23.735351 +0000 UTC m=+1545.836154499" Mar 17 09:35:27 crc kubenswrapper[4813]: I0317 09:35:27.732694 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:35:32 crc kubenswrapper[4813]: I0317 09:35:32.798859 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr" event={"ID":"df99907b-2dfb-436f-b9a0-5c62dce82672","Type":"ContainerStarted","Data":"7561ed35a7acb9cd1483fbf7a7b8898f09bf9d582c67b9747df92f979e9026f3"} Mar 17 09:35:32 crc kubenswrapper[4813]: I0317 09:35:32.827435 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr" podStartSLOduration=3.054311689 podStartE2EDuration="11.827416005s" podCreationTimestamp="2026-03-17 09:35:21 +0000 UTC" firstStartedPulling="2026-03-17 09:35:22.896143026 +0000 UTC m=+1544.996946525" lastFinishedPulling="2026-03-17 09:35:31.669247302 +0000 UTC m=+1553.770050841" observedRunningTime="2026-03-17 09:35:32.812056622 +0000 UTC m=+1554.912860121" watchObservedRunningTime="2026-03-17 09:35:32.827416005 +0000 UTC m=+1554.928219504" Mar 17 09:35:36 crc kubenswrapper[4813]: I0317 09:35:36.761671 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Mar 17 09:35:37 crc kubenswrapper[4813]: I0317 09:35:37.734894 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Mar 17 09:35:44 crc kubenswrapper[4813]: I0317 09:35:44.114501 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:35:44 crc kubenswrapper[4813]: I0317 09:35:44.115416 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:35:44 crc kubenswrapper[4813]: I0317 09:35:44.915313 4813 generic.go:334] "Generic (PLEG): container finished" podID="df99907b-2dfb-436f-b9a0-5c62dce82672" containerID="7561ed35a7acb9cd1483fbf7a7b8898f09bf9d582c67b9747df92f979e9026f3" exitCode=0 Mar 17 09:35:44 crc kubenswrapper[4813]: I0317 09:35:44.915420 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr" event={"ID":"df99907b-2dfb-436f-b9a0-5c62dce82672","Type":"ContainerDied","Data":"7561ed35a7acb9cd1483fbf7a7b8898f09bf9d582c67b9747df92f979e9026f3"} Mar 17 09:35:46 crc kubenswrapper[4813]: I0317 09:35:46.437365 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr" Mar 17 09:35:46 crc kubenswrapper[4813]: I0317 09:35:46.551326 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df99907b-2dfb-436f-b9a0-5c62dce82672-repo-setup-combined-ca-bundle\") pod \"df99907b-2dfb-436f-b9a0-5c62dce82672\" (UID: \"df99907b-2dfb-436f-b9a0-5c62dce82672\") " Mar 17 09:35:46 crc kubenswrapper[4813]: I0317 09:35:46.551750 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v6n9v\" (UniqueName: \"kubernetes.io/projected/df99907b-2dfb-436f-b9a0-5c62dce82672-kube-api-access-v6n9v\") pod \"df99907b-2dfb-436f-b9a0-5c62dce82672\" (UID: \"df99907b-2dfb-436f-b9a0-5c62dce82672\") " Mar 17 09:35:46 crc kubenswrapper[4813]: I0317 09:35:46.551864 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df99907b-2dfb-436f-b9a0-5c62dce82672-inventory\") pod \"df99907b-2dfb-436f-b9a0-5c62dce82672\" (UID: \"df99907b-2dfb-436f-b9a0-5c62dce82672\") " Mar 17 09:35:46 crc kubenswrapper[4813]: I0317 09:35:46.551954 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/df99907b-2dfb-436f-b9a0-5c62dce82672-ssh-key-openstack-edpm-ipam\") pod \"df99907b-2dfb-436f-b9a0-5c62dce82672\" (UID: \"df99907b-2dfb-436f-b9a0-5c62dce82672\") " Mar 17 09:35:46 crc kubenswrapper[4813]: I0317 09:35:46.558357 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df99907b-2dfb-436f-b9a0-5c62dce82672-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "df99907b-2dfb-436f-b9a0-5c62dce82672" (UID: "df99907b-2dfb-436f-b9a0-5c62dce82672"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:35:46 crc kubenswrapper[4813]: I0317 09:35:46.558419 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df99907b-2dfb-436f-b9a0-5c62dce82672-kube-api-access-v6n9v" (OuterVolumeSpecName: "kube-api-access-v6n9v") pod "df99907b-2dfb-436f-b9a0-5c62dce82672" (UID: "df99907b-2dfb-436f-b9a0-5c62dce82672"). InnerVolumeSpecName "kube-api-access-v6n9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:35:46 crc kubenswrapper[4813]: I0317 09:35:46.579982 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df99907b-2dfb-436f-b9a0-5c62dce82672-inventory" (OuterVolumeSpecName: "inventory") pod "df99907b-2dfb-436f-b9a0-5c62dce82672" (UID: "df99907b-2dfb-436f-b9a0-5c62dce82672"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:35:46 crc kubenswrapper[4813]: I0317 09:35:46.604031 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df99907b-2dfb-436f-b9a0-5c62dce82672-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "df99907b-2dfb-436f-b9a0-5c62dce82672" (UID: "df99907b-2dfb-436f-b9a0-5c62dce82672"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:35:46 crc kubenswrapper[4813]: I0317 09:35:46.653720 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/df99907b-2dfb-436f-b9a0-5c62dce82672-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:46 crc kubenswrapper[4813]: I0317 09:35:46.653748 4813 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df99907b-2dfb-436f-b9a0-5c62dce82672-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:46 crc kubenswrapper[4813]: I0317 09:35:46.653758 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v6n9v\" (UniqueName: \"kubernetes.io/projected/df99907b-2dfb-436f-b9a0-5c62dce82672-kube-api-access-v6n9v\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:46 crc kubenswrapper[4813]: I0317 09:35:46.653767 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/df99907b-2dfb-436f-b9a0-5c62dce82672-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:46 crc kubenswrapper[4813]: I0317 09:35:46.938212 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr" event={"ID":"df99907b-2dfb-436f-b9a0-5c62dce82672","Type":"ContainerDied","Data":"e3dd7fb5c44ecf6db1c44dbb5603dd165da18925cb8007df3c16fdc1fa5da92e"} Mar 17 09:35:46 crc kubenswrapper[4813]: I0317 09:35:46.938256 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e3dd7fb5c44ecf6db1c44dbb5603dd165da18925cb8007df3c16fdc1fa5da92e" Mar 17 09:35:46 crc kubenswrapper[4813]: I0317 09:35:46.938263 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr" Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.053429 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-5blh4"] Mar 17 09:35:47 crc kubenswrapper[4813]: E0317 09:35:47.053958 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df99907b-2dfb-436f-b9a0-5c62dce82672" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.053986 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="df99907b-2dfb-436f-b9a0-5c62dce82672" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.054307 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="df99907b-2dfb-436f-b9a0-5c62dce82672" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.055535 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5blh4" Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.059726 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hj9h4" Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.059881 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.059951 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.060205 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.068923 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-5blh4"] Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.162372 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c0a603db-d69a-47a6-8228-862e4ad835ee-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5blh4\" (UID: \"c0a603db-d69a-47a6-8228-862e4ad835ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5blh4" Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.162537 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-882nl\" (UniqueName: \"kubernetes.io/projected/c0a603db-d69a-47a6-8228-862e4ad835ee-kube-api-access-882nl\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5blh4\" (UID: \"c0a603db-d69a-47a6-8228-862e4ad835ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5blh4" Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.162624 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c0a603db-d69a-47a6-8228-862e4ad835ee-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5blh4\" (UID: \"c0a603db-d69a-47a6-8228-862e4ad835ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5blh4" Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.264470 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c0a603db-d69a-47a6-8228-862e4ad835ee-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5blh4\" (UID: \"c0a603db-d69a-47a6-8228-862e4ad835ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5blh4" Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.264716 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-882nl\" (UniqueName: \"kubernetes.io/projected/c0a603db-d69a-47a6-8228-862e4ad835ee-kube-api-access-882nl\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5blh4\" (UID: \"c0a603db-d69a-47a6-8228-862e4ad835ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5blh4" Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.264795 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c0a603db-d69a-47a6-8228-862e4ad835ee-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5blh4\" (UID: \"c0a603db-d69a-47a6-8228-862e4ad835ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5blh4" Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.272713 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c0a603db-d69a-47a6-8228-862e4ad835ee-inventory\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5blh4\" (UID: \"c0a603db-d69a-47a6-8228-862e4ad835ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5blh4" Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.272799 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c0a603db-d69a-47a6-8228-862e4ad835ee-ssh-key-openstack-edpm-ipam\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5blh4\" (UID: \"c0a603db-d69a-47a6-8228-862e4ad835ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5blh4" Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.294832 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-882nl\" (UniqueName: \"kubernetes.io/projected/c0a603db-d69a-47a6-8228-862e4ad835ee-kube-api-access-882nl\") pod \"redhat-edpm-deployment-openstack-edpm-ipam-5blh4\" (UID: \"c0a603db-d69a-47a6-8228-862e4ad835ee\") " pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5blh4" Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.388826 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5blh4" Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.914759 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.921660 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/redhat-edpm-deployment-openstack-edpm-ipam-5blh4"] Mar 17 09:35:47 crc kubenswrapper[4813]: I0317 09:35:47.948372 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5blh4" event={"ID":"c0a603db-d69a-47a6-8228-862e4ad835ee","Type":"ContainerStarted","Data":"5d5b9cc089a03ab6dcf0bc8db21d6a4fe139ed424204d852822089359d528819"} Mar 17 09:35:48 crc kubenswrapper[4813]: I0317 09:35:48.959579 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5blh4" event={"ID":"c0a603db-d69a-47a6-8228-862e4ad835ee","Type":"ContainerStarted","Data":"67b14f5c1dcb3e7016872626018d3d9ef9807eed3a1f353ee238c7ac7acd3b6f"} Mar 17 09:35:48 crc kubenswrapper[4813]: I0317 09:35:48.983889 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5blh4" podStartSLOduration=1.540660628 podStartE2EDuration="1.983868591s" podCreationTimestamp="2026-03-17 09:35:47 +0000 UTC" firstStartedPulling="2026-03-17 09:35:47.914508438 +0000 UTC m=+1570.015311927" lastFinishedPulling="2026-03-17 09:35:48.357716381 +0000 UTC m=+1570.458519890" observedRunningTime="2026-03-17 09:35:48.976888382 +0000 UTC m=+1571.077691881" watchObservedRunningTime="2026-03-17 09:35:48.983868591 +0000 UTC m=+1571.084672090" Mar 17 09:35:51 crc kubenswrapper[4813]: I0317 09:35:51.998142 4813 generic.go:334] "Generic (PLEG): container finished" podID="c0a603db-d69a-47a6-8228-862e4ad835ee" containerID="67b14f5c1dcb3e7016872626018d3d9ef9807eed3a1f353ee238c7ac7acd3b6f" exitCode=0 Mar 17 09:35:51 crc kubenswrapper[4813]: I0317 09:35:51.998214 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5blh4" event={"ID":"c0a603db-d69a-47a6-8228-862e4ad835ee","Type":"ContainerDied","Data":"67b14f5c1dcb3e7016872626018d3d9ef9807eed3a1f353ee238c7ac7acd3b6f"} Mar 17 09:35:53 crc kubenswrapper[4813]: I0317 09:35:53.518482 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5blh4" Mar 17 09:35:53 crc kubenswrapper[4813]: I0317 09:35:53.604226 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-882nl\" (UniqueName: \"kubernetes.io/projected/c0a603db-d69a-47a6-8228-862e4ad835ee-kube-api-access-882nl\") pod \"c0a603db-d69a-47a6-8228-862e4ad835ee\" (UID: \"c0a603db-d69a-47a6-8228-862e4ad835ee\") " Mar 17 09:35:53 crc kubenswrapper[4813]: I0317 09:35:53.604358 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c0a603db-d69a-47a6-8228-862e4ad835ee-inventory\") pod \"c0a603db-d69a-47a6-8228-862e4ad835ee\" (UID: \"c0a603db-d69a-47a6-8228-862e4ad835ee\") " Mar 17 09:35:53 crc kubenswrapper[4813]: I0317 09:35:53.604671 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c0a603db-d69a-47a6-8228-862e4ad835ee-ssh-key-openstack-edpm-ipam\") pod \"c0a603db-d69a-47a6-8228-862e4ad835ee\" (UID: \"c0a603db-d69a-47a6-8228-862e4ad835ee\") " Mar 17 09:35:53 crc kubenswrapper[4813]: I0317 09:35:53.610725 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0a603db-d69a-47a6-8228-862e4ad835ee-kube-api-access-882nl" (OuterVolumeSpecName: "kube-api-access-882nl") pod "c0a603db-d69a-47a6-8228-862e4ad835ee" (UID: "c0a603db-d69a-47a6-8228-862e4ad835ee"). InnerVolumeSpecName "kube-api-access-882nl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:35:53 crc kubenswrapper[4813]: I0317 09:35:53.637007 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0a603db-d69a-47a6-8228-862e4ad835ee-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c0a603db-d69a-47a6-8228-862e4ad835ee" (UID: "c0a603db-d69a-47a6-8228-862e4ad835ee"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:35:53 crc kubenswrapper[4813]: I0317 09:35:53.638698 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c0a603db-d69a-47a6-8228-862e4ad835ee-inventory" (OuterVolumeSpecName: "inventory") pod "c0a603db-d69a-47a6-8228-862e4ad835ee" (UID: "c0a603db-d69a-47a6-8228-862e4ad835ee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:35:53 crc kubenswrapper[4813]: I0317 09:35:53.707019 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c0a603db-d69a-47a6-8228-862e4ad835ee-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:53 crc kubenswrapper[4813]: I0317 09:35:53.707063 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-882nl\" (UniqueName: \"kubernetes.io/projected/c0a603db-d69a-47a6-8228-862e4ad835ee-kube-api-access-882nl\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:53 crc kubenswrapper[4813]: I0317 09:35:53.707077 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c0a603db-d69a-47a6-8228-862e4ad835ee-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.025169 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5blh4" event={"ID":"c0a603db-d69a-47a6-8228-862e4ad835ee","Type":"ContainerDied","Data":"5d5b9cc089a03ab6dcf0bc8db21d6a4fe139ed424204d852822089359d528819"} Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.025250 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d5b9cc089a03ab6dcf0bc8db21d6a4fe139ed424204d852822089359d528819" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.025356 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/redhat-edpm-deployment-openstack-edpm-ipam-5blh4" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.129141 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg"] Mar 17 09:35:54 crc kubenswrapper[4813]: E0317 09:35:54.129528 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0a603db-d69a-47a6-8228-862e4ad835ee" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.129542 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0a603db-d69a-47a6-8228-862e4ad835ee" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.129742 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0a603db-d69a-47a6-8228-862e4ad835ee" containerName="redhat-edpm-deployment-openstack-edpm-ipam" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.130312 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.133325 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.133524 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.133666 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hj9h4" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.134572 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.138912 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg"] Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.219791 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f9166d03-0282-47e6-a80b-5b03f4183f62-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg\" (UID: \"f9166d03-0282-47e6-a80b-5b03f4183f62\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.219950 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsspg\" (UniqueName: \"kubernetes.io/projected/f9166d03-0282-47e6-a80b-5b03f4183f62-kube-api-access-xsspg\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg\" (UID: \"f9166d03-0282-47e6-a80b-5b03f4183f62\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.219987 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9166d03-0282-47e6-a80b-5b03f4183f62-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg\" (UID: \"f9166d03-0282-47e6-a80b-5b03f4183f62\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.220031 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9166d03-0282-47e6-a80b-5b03f4183f62-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg\" (UID: \"f9166d03-0282-47e6-a80b-5b03f4183f62\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.321862 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsspg\" (UniqueName: \"kubernetes.io/projected/f9166d03-0282-47e6-a80b-5b03f4183f62-kube-api-access-xsspg\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg\" (UID: \"f9166d03-0282-47e6-a80b-5b03f4183f62\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.322305 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9166d03-0282-47e6-a80b-5b03f4183f62-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg\" (UID: \"f9166d03-0282-47e6-a80b-5b03f4183f62\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.322413 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9166d03-0282-47e6-a80b-5b03f4183f62-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg\" (UID: \"f9166d03-0282-47e6-a80b-5b03f4183f62\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.322531 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f9166d03-0282-47e6-a80b-5b03f4183f62-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg\" (UID: \"f9166d03-0282-47e6-a80b-5b03f4183f62\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.327660 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9166d03-0282-47e6-a80b-5b03f4183f62-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg\" (UID: \"f9166d03-0282-47e6-a80b-5b03f4183f62\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.328715 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f9166d03-0282-47e6-a80b-5b03f4183f62-ssh-key-openstack-edpm-ipam\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg\" (UID: \"f9166d03-0282-47e6-a80b-5b03f4183f62\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.331539 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9166d03-0282-47e6-a80b-5b03f4183f62-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg\" (UID: \"f9166d03-0282-47e6-a80b-5b03f4183f62\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.354445 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsspg\" (UniqueName: \"kubernetes.io/projected/f9166d03-0282-47e6-a80b-5b03f4183f62-kube-api-access-xsspg\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg\" (UID: \"f9166d03-0282-47e6-a80b-5b03f4183f62\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg" Mar 17 09:35:54 crc kubenswrapper[4813]: I0317 09:35:54.458035 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg" Mar 17 09:35:55 crc kubenswrapper[4813]: I0317 09:35:55.116732 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg"] Mar 17 09:35:55 crc kubenswrapper[4813]: W0317 09:35:55.122901 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9166d03_0282_47e6_a80b_5b03f4183f62.slice/crio-c0863577e4b48002b17ce662851a175f8082223e553bc8eabcd8105d36664422 WatchSource:0}: Error finding container c0863577e4b48002b17ce662851a175f8082223e553bc8eabcd8105d36664422: Status 404 returned error can't find the container with id c0863577e4b48002b17ce662851a175f8082223e553bc8eabcd8105d36664422 Mar 17 09:35:56 crc kubenswrapper[4813]: I0317 09:35:56.052939 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg" event={"ID":"f9166d03-0282-47e6-a80b-5b03f4183f62","Type":"ContainerStarted","Data":"33371410433e777102a5b0239b957318241f07806fd68a71b07d16102d757f9f"} Mar 17 09:35:56 crc kubenswrapper[4813]: I0317 09:35:56.053334 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg" event={"ID":"f9166d03-0282-47e6-a80b-5b03f4183f62","Type":"ContainerStarted","Data":"c0863577e4b48002b17ce662851a175f8082223e553bc8eabcd8105d36664422"} Mar 17 09:35:56 crc kubenswrapper[4813]: I0317 09:35:56.109470 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg" podStartSLOduration=1.738311519 podStartE2EDuration="2.109435318s" podCreationTimestamp="2026-03-17 09:35:54 +0000 UTC" firstStartedPulling="2026-03-17 09:35:55.126528 +0000 UTC m=+1577.227331539" lastFinishedPulling="2026-03-17 09:35:55.497651839 +0000 UTC m=+1577.598455338" observedRunningTime="2026-03-17 09:35:56.07765992 +0000 UTC m=+1578.178463459" watchObservedRunningTime="2026-03-17 09:35:56.109435318 +0000 UTC m=+1578.210238857" Mar 17 09:36:00 crc kubenswrapper[4813]: I0317 09:36:00.171912 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562336-c6fds"] Mar 17 09:36:00 crc kubenswrapper[4813]: I0317 09:36:00.174455 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562336-c6fds" Mar 17 09:36:00 crc kubenswrapper[4813]: I0317 09:36:00.179383 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:36:00 crc kubenswrapper[4813]: I0317 09:36:00.184750 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:36:00 crc kubenswrapper[4813]: I0317 09:36:00.194933 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:36:00 crc kubenswrapper[4813]: I0317 09:36:00.206090 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562336-c6fds"] Mar 17 09:36:00 crc kubenswrapper[4813]: I0317 09:36:00.259702 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7msp\" (UniqueName: \"kubernetes.io/projected/0afc91f2-4193-40e2-9654-65a1b31326c3-kube-api-access-z7msp\") pod \"auto-csr-approver-29562336-c6fds\" (UID: \"0afc91f2-4193-40e2-9654-65a1b31326c3\") " pod="openshift-infra/auto-csr-approver-29562336-c6fds" Mar 17 09:36:00 crc kubenswrapper[4813]: I0317 09:36:00.362688 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7msp\" (UniqueName: \"kubernetes.io/projected/0afc91f2-4193-40e2-9654-65a1b31326c3-kube-api-access-z7msp\") pod \"auto-csr-approver-29562336-c6fds\" (UID: \"0afc91f2-4193-40e2-9654-65a1b31326c3\") " pod="openshift-infra/auto-csr-approver-29562336-c6fds" Mar 17 09:36:00 crc kubenswrapper[4813]: I0317 09:36:00.395396 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7msp\" (UniqueName: \"kubernetes.io/projected/0afc91f2-4193-40e2-9654-65a1b31326c3-kube-api-access-z7msp\") pod \"auto-csr-approver-29562336-c6fds\" (UID: \"0afc91f2-4193-40e2-9654-65a1b31326c3\") " pod="openshift-infra/auto-csr-approver-29562336-c6fds" Mar 17 09:36:00 crc kubenswrapper[4813]: I0317 09:36:00.509981 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562336-c6fds" Mar 17 09:36:00 crc kubenswrapper[4813]: I0317 09:36:00.994238 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562336-c6fds"] Mar 17 09:36:01 crc kubenswrapper[4813]: I0317 09:36:01.113079 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562336-c6fds" event={"ID":"0afc91f2-4193-40e2-9654-65a1b31326c3","Type":"ContainerStarted","Data":"f751b21e417c28f3147015e91fcdc129e4583d18c9007fce015b3f4e3fd7e8ce"} Mar 17 09:36:03 crc kubenswrapper[4813]: I0317 09:36:03.140987 4813 generic.go:334] "Generic (PLEG): container finished" podID="0afc91f2-4193-40e2-9654-65a1b31326c3" containerID="d58cc6f0d9a5c4bb85d3259485359f7fe7e20d5f74705fe50c90a3f040deae9c" exitCode=0 Mar 17 09:36:03 crc kubenswrapper[4813]: I0317 09:36:03.141305 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562336-c6fds" event={"ID":"0afc91f2-4193-40e2-9654-65a1b31326c3","Type":"ContainerDied","Data":"d58cc6f0d9a5c4bb85d3259485359f7fe7e20d5f74705fe50c90a3f040deae9c"} Mar 17 09:36:04 crc kubenswrapper[4813]: I0317 09:36:04.542796 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562336-c6fds" Mar 17 09:36:04 crc kubenswrapper[4813]: I0317 09:36:04.652022 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7msp\" (UniqueName: \"kubernetes.io/projected/0afc91f2-4193-40e2-9654-65a1b31326c3-kube-api-access-z7msp\") pod \"0afc91f2-4193-40e2-9654-65a1b31326c3\" (UID: \"0afc91f2-4193-40e2-9654-65a1b31326c3\") " Mar 17 09:36:04 crc kubenswrapper[4813]: I0317 09:36:04.660945 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0afc91f2-4193-40e2-9654-65a1b31326c3-kube-api-access-z7msp" (OuterVolumeSpecName: "kube-api-access-z7msp") pod "0afc91f2-4193-40e2-9654-65a1b31326c3" (UID: "0afc91f2-4193-40e2-9654-65a1b31326c3"). InnerVolumeSpecName "kube-api-access-z7msp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:36:04 crc kubenswrapper[4813]: I0317 09:36:04.754376 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7msp\" (UniqueName: \"kubernetes.io/projected/0afc91f2-4193-40e2-9654-65a1b31326c3-kube-api-access-z7msp\") on node \"crc\" DevicePath \"\"" Mar 17 09:36:05 crc kubenswrapper[4813]: I0317 09:36:05.165494 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562336-c6fds" event={"ID":"0afc91f2-4193-40e2-9654-65a1b31326c3","Type":"ContainerDied","Data":"f751b21e417c28f3147015e91fcdc129e4583d18c9007fce015b3f4e3fd7e8ce"} Mar 17 09:36:05 crc kubenswrapper[4813]: I0317 09:36:05.165542 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f751b21e417c28f3147015e91fcdc129e4583d18c9007fce015b3f4e3fd7e8ce" Mar 17 09:36:05 crc kubenswrapper[4813]: I0317 09:36:05.165564 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562336-c6fds" Mar 17 09:36:05 crc kubenswrapper[4813]: I0317 09:36:05.630134 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562330-b5zg5"] Mar 17 09:36:05 crc kubenswrapper[4813]: I0317 09:36:05.645591 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562330-b5zg5"] Mar 17 09:36:06 crc kubenswrapper[4813]: I0317 09:36:06.740691 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="339d3b99-3aa5-4c3b-99b2-997de272c26c" path="/var/lib/kubelet/pods/339d3b99-3aa5-4c3b-99b2-997de272c26c/volumes" Mar 17 09:36:14 crc kubenswrapper[4813]: I0317 09:36:14.113590 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:36:14 crc kubenswrapper[4813]: I0317 09:36:14.114124 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:36:44 crc kubenswrapper[4813]: I0317 09:36:44.114694 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:36:44 crc kubenswrapper[4813]: I0317 09:36:44.115273 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:36:44 crc kubenswrapper[4813]: I0317 09:36:44.115356 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:36:44 crc kubenswrapper[4813]: I0317 09:36:44.116517 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368"} pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 09:36:44 crc kubenswrapper[4813]: I0317 09:36:44.116656 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" containerID="cri-o://cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" gracePeriod=600 Mar 17 09:36:44 crc kubenswrapper[4813]: E0317 09:36:44.253548 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:36:45 crc kubenswrapper[4813]: I0317 09:36:45.209719 4813 generic.go:334] "Generic (PLEG): container finished" podID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" exitCode=0 Mar 17 09:36:45 crc kubenswrapper[4813]: I0317 09:36:45.209798 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerDied","Data":"cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368"} Mar 17 09:36:45 crc kubenswrapper[4813]: I0317 09:36:45.210130 4813 scope.go:117] "RemoveContainer" containerID="4ba33e308a10bfb001c8289b81c4de34337bf84d52f75bd8b871b725edd0f79b" Mar 17 09:36:45 crc kubenswrapper[4813]: I0317 09:36:45.210533 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:36:45 crc kubenswrapper[4813]: E0317 09:36:45.210959 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:36:45 crc kubenswrapper[4813]: I0317 09:36:45.597521 4813 scope.go:117] "RemoveContainer" containerID="7b352d6cb96798d6f939eb7dc8ae0edd91a09eaeac9d1b8058882634055754d6" Mar 17 09:36:45 crc kubenswrapper[4813]: I0317 09:36:45.671477 4813 scope.go:117] "RemoveContainer" containerID="4447f4ba14b78f065ba1ae8a1456f1f1fd8401676d81db53dfe718ae82ef145d" Mar 17 09:36:45 crc kubenswrapper[4813]: I0317 09:36:45.700896 4813 scope.go:117] "RemoveContainer" containerID="69873978b52fa34c4f053a1af3b3d6f826706713d0fea838103a273204174cbe" Mar 17 09:36:45 crc kubenswrapper[4813]: I0317 09:36:45.764221 4813 scope.go:117] "RemoveContainer" containerID="cc5738221e82f457b14cb9dfb0e94faa09e5204159938b7bebabc755f0a806d5" Mar 17 09:36:59 crc kubenswrapper[4813]: I0317 09:36:59.731552 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:36:59 crc kubenswrapper[4813]: E0317 09:36:59.732403 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:37:12 crc kubenswrapper[4813]: I0317 09:37:12.730480 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:37:12 crc kubenswrapper[4813]: E0317 09:37:12.731135 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:37:24 crc kubenswrapper[4813]: I0317 09:37:24.730554 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:37:24 crc kubenswrapper[4813]: E0317 09:37:24.731435 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:37:37 crc kubenswrapper[4813]: I0317 09:37:37.731186 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:37:37 crc kubenswrapper[4813]: E0317 09:37:37.732204 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:37:45 crc kubenswrapper[4813]: I0317 09:37:45.901659 4813 scope.go:117] "RemoveContainer" containerID="a23facb1f8d187a1f432ea018fb2ca53da0f6eacfe0d1b993f572c13e0723bf0" Mar 17 09:37:45 crc kubenswrapper[4813]: I0317 09:37:45.940203 4813 scope.go:117] "RemoveContainer" containerID="0915e99e93df5119ef52f738f246dcf25eb4dd780225e4b1f659ae48abeb232b" Mar 17 09:37:45 crc kubenswrapper[4813]: I0317 09:37:45.990971 4813 scope.go:117] "RemoveContainer" containerID="c54d5cf6483e86ddc4f03258cfc9647987f653d9df2e0d792d725827ee504ea8" Mar 17 09:37:51 crc kubenswrapper[4813]: I0317 09:37:51.731445 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:37:51 crc kubenswrapper[4813]: E0317 09:37:51.732258 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:38:00 crc kubenswrapper[4813]: I0317 09:38:00.168418 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562338-qjcvb"] Mar 17 09:38:00 crc kubenswrapper[4813]: E0317 09:38:00.170836 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0afc91f2-4193-40e2-9654-65a1b31326c3" containerName="oc" Mar 17 09:38:00 crc kubenswrapper[4813]: I0317 09:38:00.170875 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0afc91f2-4193-40e2-9654-65a1b31326c3" containerName="oc" Mar 17 09:38:00 crc kubenswrapper[4813]: I0317 09:38:00.171366 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0afc91f2-4193-40e2-9654-65a1b31326c3" containerName="oc" Mar 17 09:38:00 crc kubenswrapper[4813]: I0317 09:38:00.172864 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562338-qjcvb" Mar 17 09:38:00 crc kubenswrapper[4813]: I0317 09:38:00.176544 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:38:00 crc kubenswrapper[4813]: I0317 09:38:00.178797 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:38:00 crc kubenswrapper[4813]: I0317 09:38:00.179370 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:38:00 crc kubenswrapper[4813]: I0317 09:38:00.181664 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562338-qjcvb"] Mar 17 09:38:00 crc kubenswrapper[4813]: I0317 09:38:00.254307 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7zk67\" (UniqueName: \"kubernetes.io/projected/6ef5582f-769e-40ee-bcf1-56d5509d19e6-kube-api-access-7zk67\") pod \"auto-csr-approver-29562338-qjcvb\" (UID: \"6ef5582f-769e-40ee-bcf1-56d5509d19e6\") " pod="openshift-infra/auto-csr-approver-29562338-qjcvb" Mar 17 09:38:00 crc kubenswrapper[4813]: I0317 09:38:00.356377 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7zk67\" (UniqueName: \"kubernetes.io/projected/6ef5582f-769e-40ee-bcf1-56d5509d19e6-kube-api-access-7zk67\") pod \"auto-csr-approver-29562338-qjcvb\" (UID: \"6ef5582f-769e-40ee-bcf1-56d5509d19e6\") " pod="openshift-infra/auto-csr-approver-29562338-qjcvb" Mar 17 09:38:00 crc kubenswrapper[4813]: I0317 09:38:00.380507 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7zk67\" (UniqueName: \"kubernetes.io/projected/6ef5582f-769e-40ee-bcf1-56d5509d19e6-kube-api-access-7zk67\") pod \"auto-csr-approver-29562338-qjcvb\" (UID: \"6ef5582f-769e-40ee-bcf1-56d5509d19e6\") " pod="openshift-infra/auto-csr-approver-29562338-qjcvb" Mar 17 09:38:00 crc kubenswrapper[4813]: I0317 09:38:00.500685 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562338-qjcvb" Mar 17 09:38:01 crc kubenswrapper[4813]: I0317 09:38:01.002838 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562338-qjcvb"] Mar 17 09:38:01 crc kubenswrapper[4813]: W0317 09:38:01.008462 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ef5582f_769e_40ee_bcf1_56d5509d19e6.slice/crio-7ab3f12962c19e8db77800b93ef26347fb655077e26b7e2163b37c4529d9ecdb WatchSource:0}: Error finding container 7ab3f12962c19e8db77800b93ef26347fb655077e26b7e2163b37c4529d9ecdb: Status 404 returned error can't find the container with id 7ab3f12962c19e8db77800b93ef26347fb655077e26b7e2163b37c4529d9ecdb Mar 17 09:38:01 crc kubenswrapper[4813]: I0317 09:38:01.078756 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562338-qjcvb" event={"ID":"6ef5582f-769e-40ee-bcf1-56d5509d19e6","Type":"ContainerStarted","Data":"7ab3f12962c19e8db77800b93ef26347fb655077e26b7e2163b37c4529d9ecdb"} Mar 17 09:38:03 crc kubenswrapper[4813]: I0317 09:38:03.115370 4813 generic.go:334] "Generic (PLEG): container finished" podID="6ef5582f-769e-40ee-bcf1-56d5509d19e6" containerID="f129d5928c7aa0b103d13c302017e8fb28643b797037fdeccd6472f22d221548" exitCode=0 Mar 17 09:38:03 crc kubenswrapper[4813]: I0317 09:38:03.115428 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562338-qjcvb" event={"ID":"6ef5582f-769e-40ee-bcf1-56d5509d19e6","Type":"ContainerDied","Data":"f129d5928c7aa0b103d13c302017e8fb28643b797037fdeccd6472f22d221548"} Mar 17 09:38:04 crc kubenswrapper[4813]: I0317 09:38:04.552022 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562338-qjcvb" Mar 17 09:38:04 crc kubenswrapper[4813]: I0317 09:38:04.741881 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7zk67\" (UniqueName: \"kubernetes.io/projected/6ef5582f-769e-40ee-bcf1-56d5509d19e6-kube-api-access-7zk67\") pod \"6ef5582f-769e-40ee-bcf1-56d5509d19e6\" (UID: \"6ef5582f-769e-40ee-bcf1-56d5509d19e6\") " Mar 17 09:38:04 crc kubenswrapper[4813]: I0317 09:38:04.747223 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ef5582f-769e-40ee-bcf1-56d5509d19e6-kube-api-access-7zk67" (OuterVolumeSpecName: "kube-api-access-7zk67") pod "6ef5582f-769e-40ee-bcf1-56d5509d19e6" (UID: "6ef5582f-769e-40ee-bcf1-56d5509d19e6"). InnerVolumeSpecName "kube-api-access-7zk67". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:38:04 crc kubenswrapper[4813]: I0317 09:38:04.844648 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7zk67\" (UniqueName: \"kubernetes.io/projected/6ef5582f-769e-40ee-bcf1-56d5509d19e6-kube-api-access-7zk67\") on node \"crc\" DevicePath \"\"" Mar 17 09:38:05 crc kubenswrapper[4813]: I0317 09:38:05.138564 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562338-qjcvb" event={"ID":"6ef5582f-769e-40ee-bcf1-56d5509d19e6","Type":"ContainerDied","Data":"7ab3f12962c19e8db77800b93ef26347fb655077e26b7e2163b37c4529d9ecdb"} Mar 17 09:38:05 crc kubenswrapper[4813]: I0317 09:38:05.138882 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ab3f12962c19e8db77800b93ef26347fb655077e26b7e2163b37c4529d9ecdb" Mar 17 09:38:05 crc kubenswrapper[4813]: I0317 09:38:05.138730 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562338-qjcvb" Mar 17 09:38:05 crc kubenswrapper[4813]: I0317 09:38:05.658923 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562332-p5phc"] Mar 17 09:38:05 crc kubenswrapper[4813]: I0317 09:38:05.673081 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562332-p5phc"] Mar 17 09:38:06 crc kubenswrapper[4813]: I0317 09:38:06.731936 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:38:06 crc kubenswrapper[4813]: E0317 09:38:06.732196 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:38:06 crc kubenswrapper[4813]: I0317 09:38:06.749588 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cf3e585-9c98-4e66-b577-59ddd1327ca2" path="/var/lib/kubelet/pods/4cf3e585-9c98-4e66-b577-59ddd1327ca2/volumes" Mar 17 09:38:21 crc kubenswrapper[4813]: I0317 09:38:21.731069 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:38:21 crc kubenswrapper[4813]: E0317 09:38:21.732135 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:38:34 crc kubenswrapper[4813]: I0317 09:38:34.731752 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:38:34 crc kubenswrapper[4813]: E0317 09:38:34.733169 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:38:46 crc kubenswrapper[4813]: I0317 09:38:46.115530 4813 scope.go:117] "RemoveContainer" containerID="dcdf36ad4ecc1c39dbd97f7dff700cc7e19e7a9ed7577401ceb8bb9ee4a3e43d" Mar 17 09:38:46 crc kubenswrapper[4813]: I0317 09:38:46.667011 4813 generic.go:334] "Generic (PLEG): container finished" podID="f9166d03-0282-47e6-a80b-5b03f4183f62" containerID="33371410433e777102a5b0239b957318241f07806fd68a71b07d16102d757f9f" exitCode=0 Mar 17 09:38:46 crc kubenswrapper[4813]: I0317 09:38:46.667142 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg" event={"ID":"f9166d03-0282-47e6-a80b-5b03f4183f62","Type":"ContainerDied","Data":"33371410433e777102a5b0239b957318241f07806fd68a71b07d16102d757f9f"} Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.167607 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.305310 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9166d03-0282-47e6-a80b-5b03f4183f62-inventory\") pod \"f9166d03-0282-47e6-a80b-5b03f4183f62\" (UID: \"f9166d03-0282-47e6-a80b-5b03f4183f62\") " Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.305742 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsspg\" (UniqueName: \"kubernetes.io/projected/f9166d03-0282-47e6-a80b-5b03f4183f62-kube-api-access-xsspg\") pod \"f9166d03-0282-47e6-a80b-5b03f4183f62\" (UID: \"f9166d03-0282-47e6-a80b-5b03f4183f62\") " Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.306242 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9166d03-0282-47e6-a80b-5b03f4183f62-bootstrap-combined-ca-bundle\") pod \"f9166d03-0282-47e6-a80b-5b03f4183f62\" (UID: \"f9166d03-0282-47e6-a80b-5b03f4183f62\") " Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.306367 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f9166d03-0282-47e6-a80b-5b03f4183f62-ssh-key-openstack-edpm-ipam\") pod \"f9166d03-0282-47e6-a80b-5b03f4183f62\" (UID: \"f9166d03-0282-47e6-a80b-5b03f4183f62\") " Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.314492 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9166d03-0282-47e6-a80b-5b03f4183f62-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "f9166d03-0282-47e6-a80b-5b03f4183f62" (UID: "f9166d03-0282-47e6-a80b-5b03f4183f62"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.316247 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9166d03-0282-47e6-a80b-5b03f4183f62-kube-api-access-xsspg" (OuterVolumeSpecName: "kube-api-access-xsspg") pod "f9166d03-0282-47e6-a80b-5b03f4183f62" (UID: "f9166d03-0282-47e6-a80b-5b03f4183f62"). InnerVolumeSpecName "kube-api-access-xsspg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.353265 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9166d03-0282-47e6-a80b-5b03f4183f62-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "f9166d03-0282-47e6-a80b-5b03f4183f62" (UID: "f9166d03-0282-47e6-a80b-5b03f4183f62"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.368864 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f9166d03-0282-47e6-a80b-5b03f4183f62-inventory" (OuterVolumeSpecName: "inventory") pod "f9166d03-0282-47e6-a80b-5b03f4183f62" (UID: "f9166d03-0282-47e6-a80b-5b03f4183f62"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.408954 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsspg\" (UniqueName: \"kubernetes.io/projected/f9166d03-0282-47e6-a80b-5b03f4183f62-kube-api-access-xsspg\") on node \"crc\" DevicePath \"\"" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.409006 4813 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f9166d03-0282-47e6-a80b-5b03f4183f62-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.409027 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/f9166d03-0282-47e6-a80b-5b03f4183f62-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.409079 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f9166d03-0282-47e6-a80b-5b03f4183f62-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.690094 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg" event={"ID":"f9166d03-0282-47e6-a80b-5b03f4183f62","Type":"ContainerDied","Data":"c0863577e4b48002b17ce662851a175f8082223e553bc8eabcd8105d36664422"} Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.690142 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0863577e4b48002b17ce662851a175f8082223e553bc8eabcd8105d36664422" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.690177 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.751882 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:38:48 crc kubenswrapper[4813]: E0317 09:38:48.752453 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.798287 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46np8"] Mar 17 09:38:48 crc kubenswrapper[4813]: E0317 09:38:48.799060 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ef5582f-769e-40ee-bcf1-56d5509d19e6" containerName="oc" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.799092 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ef5582f-769e-40ee-bcf1-56d5509d19e6" containerName="oc" Mar 17 09:38:48 crc kubenswrapper[4813]: E0317 09:38:48.799127 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9166d03-0282-47e6-a80b-5b03f4183f62" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.799141 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9166d03-0282-47e6-a80b-5b03f4183f62" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.799344 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ef5582f-769e-40ee-bcf1-56d5509d19e6" containerName="oc" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.799375 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9166d03-0282-47e6-a80b-5b03f4183f62" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.800306 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46np8" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.802419 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.802490 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.802628 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hj9h4" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.803430 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.820136 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46np8"] Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.952453 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36536921-03a1-45dd-93b3-1d06e9c3adca-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-46np8\" (UID: \"36536921-03a1-45dd-93b3-1d06e9c3adca\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46np8" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.952510 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/36536921-03a1-45dd-93b3-1d06e9c3adca-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-46np8\" (UID: \"36536921-03a1-45dd-93b3-1d06e9c3adca\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46np8" Mar 17 09:38:48 crc kubenswrapper[4813]: I0317 09:38:48.952580 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmdk7\" (UniqueName: \"kubernetes.io/projected/36536921-03a1-45dd-93b3-1d06e9c3adca-kube-api-access-qmdk7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-46np8\" (UID: \"36536921-03a1-45dd-93b3-1d06e9c3adca\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46np8" Mar 17 09:38:49 crc kubenswrapper[4813]: I0317 09:38:49.054102 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36536921-03a1-45dd-93b3-1d06e9c3adca-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-46np8\" (UID: \"36536921-03a1-45dd-93b3-1d06e9c3adca\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46np8" Mar 17 09:38:49 crc kubenswrapper[4813]: I0317 09:38:49.054155 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/36536921-03a1-45dd-93b3-1d06e9c3adca-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-46np8\" (UID: \"36536921-03a1-45dd-93b3-1d06e9c3adca\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46np8" Mar 17 09:38:49 crc kubenswrapper[4813]: I0317 09:38:49.054212 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmdk7\" (UniqueName: \"kubernetes.io/projected/36536921-03a1-45dd-93b3-1d06e9c3adca-kube-api-access-qmdk7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-46np8\" (UID: \"36536921-03a1-45dd-93b3-1d06e9c3adca\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46np8" Mar 17 09:38:49 crc kubenswrapper[4813]: I0317 09:38:49.063261 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36536921-03a1-45dd-93b3-1d06e9c3adca-inventory\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-46np8\" (UID: \"36536921-03a1-45dd-93b3-1d06e9c3adca\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46np8" Mar 17 09:38:49 crc kubenswrapper[4813]: I0317 09:38:49.064194 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/36536921-03a1-45dd-93b3-1d06e9c3adca-ssh-key-openstack-edpm-ipam\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-46np8\" (UID: \"36536921-03a1-45dd-93b3-1d06e9c3adca\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46np8" Mar 17 09:38:49 crc kubenswrapper[4813]: I0317 09:38:49.070764 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmdk7\" (UniqueName: \"kubernetes.io/projected/36536921-03a1-45dd-93b3-1d06e9c3adca-kube-api-access-qmdk7\") pod \"download-cache-edpm-deployment-openstack-edpm-ipam-46np8\" (UID: \"36536921-03a1-45dd-93b3-1d06e9c3adca\") " pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46np8" Mar 17 09:38:49 crc kubenswrapper[4813]: I0317 09:38:49.117206 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46np8" Mar 17 09:38:49 crc kubenswrapper[4813]: I0317 09:38:49.632243 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46np8"] Mar 17 09:38:49 crc kubenswrapper[4813]: I0317 09:38:49.701618 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46np8" event={"ID":"36536921-03a1-45dd-93b3-1d06e9c3adca","Type":"ContainerStarted","Data":"35dda0fcf8586e0e107cce41b0e526075d81575b40b15b61dd5532b56201e720"} Mar 17 09:38:50 crc kubenswrapper[4813]: I0317 09:38:50.715180 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46np8" event={"ID":"36536921-03a1-45dd-93b3-1d06e9c3adca","Type":"ContainerStarted","Data":"fe26fb49a47dd7bdf5785d4e92a4e5c9c609ee0f77b9eafd13e7462bde10f26b"} Mar 17 09:38:50 crc kubenswrapper[4813]: I0317 09:38:50.746211 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46np8" podStartSLOduration=2.252734956 podStartE2EDuration="2.746187666s" podCreationTimestamp="2026-03-17 09:38:48 +0000 UTC" firstStartedPulling="2026-03-17 09:38:49.639214952 +0000 UTC m=+1751.740018461" lastFinishedPulling="2026-03-17 09:38:50.132667662 +0000 UTC m=+1752.233471171" observedRunningTime="2026-03-17 09:38:50.743865466 +0000 UTC m=+1752.844669005" watchObservedRunningTime="2026-03-17 09:38:50.746187666 +0000 UTC m=+1752.846991205" Mar 17 09:39:03 crc kubenswrapper[4813]: I0317 09:39:03.730720 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:39:03 crc kubenswrapper[4813]: E0317 09:39:03.731717 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:39:16 crc kubenswrapper[4813]: I0317 09:39:16.731483 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:39:16 crc kubenswrapper[4813]: E0317 09:39:16.732240 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:39:27 crc kubenswrapper[4813]: I0317 09:39:27.730904 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:39:27 crc kubenswrapper[4813]: E0317 09:39:27.731996 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:39:39 crc kubenswrapper[4813]: I0317 09:39:39.731484 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:39:39 crc kubenswrapper[4813]: E0317 09:39:39.732432 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:39:46 crc kubenswrapper[4813]: I0317 09:39:46.216146 4813 scope.go:117] "RemoveContainer" containerID="987f42380506732cb7bccd1f21a56d81e217d17ebf5e5f8bb56bd3397368fdac" Mar 17 09:39:46 crc kubenswrapper[4813]: I0317 09:39:46.253914 4813 scope.go:117] "RemoveContainer" containerID="0b9a534ff3bf6cd920494816e2d55e49a081e9a72e266ef15de661be98e603b2" Mar 17 09:39:46 crc kubenswrapper[4813]: I0317 09:39:46.273068 4813 scope.go:117] "RemoveContainer" containerID="a2de365165a9e93f392da0a42a918ce0e6cf111a582c5b174f55c44c7d966791" Mar 17 09:39:46 crc kubenswrapper[4813]: I0317 09:39:46.299053 4813 scope.go:117] "RemoveContainer" containerID="a44bfd63612dfab25234155cad9e67aa869107fcf44e75fc72bab3b9bef5eb28" Mar 17 09:39:53 crc kubenswrapper[4813]: I0317 09:39:53.731625 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:39:53 crc kubenswrapper[4813]: E0317 09:39:53.732992 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:39:54 crc kubenswrapper[4813]: I0317 09:39:54.069965 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-6ee7-account-create-update-ffh6m"] Mar 17 09:39:54 crc kubenswrapper[4813]: I0317 09:39:54.084344 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-6ee7-account-create-update-ffh6m"] Mar 17 09:39:54 crc kubenswrapper[4813]: I0317 09:39:54.742432 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35511914-5b8f-45e4-87cb-cd174e467dcc" path="/var/lib/kubelet/pods/35511914-5b8f-45e4-87cb-cd174e467dcc/volumes" Mar 17 09:40:00 crc kubenswrapper[4813]: I0317 09:40:00.164195 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562340-cw8rp"] Mar 17 09:40:00 crc kubenswrapper[4813]: I0317 09:40:00.168061 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562340-cw8rp" Mar 17 09:40:00 crc kubenswrapper[4813]: I0317 09:40:00.170963 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:40:00 crc kubenswrapper[4813]: I0317 09:40:00.171096 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:40:00 crc kubenswrapper[4813]: I0317 09:40:00.172756 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:40:00 crc kubenswrapper[4813]: I0317 09:40:00.181511 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562340-cw8rp"] Mar 17 09:40:00 crc kubenswrapper[4813]: I0317 09:40:00.275883 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4gqnh\" (UniqueName: \"kubernetes.io/projected/8b5d72c0-ddd6-4c41-8845-7052c340a046-kube-api-access-4gqnh\") pod \"auto-csr-approver-29562340-cw8rp\" (UID: \"8b5d72c0-ddd6-4c41-8845-7052c340a046\") " pod="openshift-infra/auto-csr-approver-29562340-cw8rp" Mar 17 09:40:00 crc kubenswrapper[4813]: I0317 09:40:00.377262 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4gqnh\" (UniqueName: \"kubernetes.io/projected/8b5d72c0-ddd6-4c41-8845-7052c340a046-kube-api-access-4gqnh\") pod \"auto-csr-approver-29562340-cw8rp\" (UID: \"8b5d72c0-ddd6-4c41-8845-7052c340a046\") " pod="openshift-infra/auto-csr-approver-29562340-cw8rp" Mar 17 09:40:00 crc kubenswrapper[4813]: I0317 09:40:00.398906 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4gqnh\" (UniqueName: \"kubernetes.io/projected/8b5d72c0-ddd6-4c41-8845-7052c340a046-kube-api-access-4gqnh\") pod \"auto-csr-approver-29562340-cw8rp\" (UID: \"8b5d72c0-ddd6-4c41-8845-7052c340a046\") " pod="openshift-infra/auto-csr-approver-29562340-cw8rp" Mar 17 09:40:00 crc kubenswrapper[4813]: I0317 09:40:00.496935 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562340-cw8rp" Mar 17 09:40:01 crc kubenswrapper[4813]: I0317 09:40:01.056447 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562340-cw8rp"] Mar 17 09:40:01 crc kubenswrapper[4813]: I0317 09:40:01.191936 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562340-cw8rp" event={"ID":"8b5d72c0-ddd6-4c41-8845-7052c340a046","Type":"ContainerStarted","Data":"93e5ab76653f205dee577192a71d2fbce73d9e52923d5d89bef31410ef717238"} Mar 17 09:40:02 crc kubenswrapper[4813]: I0317 09:40:02.053163 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-efef-account-create-update-wplm2"] Mar 17 09:40:02 crc kubenswrapper[4813]: I0317 09:40:02.069480 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-46tzl"] Mar 17 09:40:02 crc kubenswrapper[4813]: I0317 09:40:02.080018 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-efef-account-create-update-wplm2"] Mar 17 09:40:02 crc kubenswrapper[4813]: I0317 09:40:02.090248 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-46tzl"] Mar 17 09:40:02 crc kubenswrapper[4813]: I0317 09:40:02.754581 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f7359df-58d8-4c25-91b5-d262b088d46e" path="/var/lib/kubelet/pods/0f7359df-58d8-4c25-91b5-d262b088d46e/volumes" Mar 17 09:40:02 crc kubenswrapper[4813]: I0317 09:40:02.757701 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b380124-90e1-48e8-a3f1-fca49c1adf07" path="/var/lib/kubelet/pods/6b380124-90e1-48e8-a3f1-fca49c1adf07/volumes" Mar 17 09:40:03 crc kubenswrapper[4813]: I0317 09:40:03.218836 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562340-cw8rp" event={"ID":"8b5d72c0-ddd6-4c41-8845-7052c340a046","Type":"ContainerStarted","Data":"09b1e1e803fe358a61aa5bfed604dbac24f44be69505e18ae4d8b75ed810f988"} Mar 17 09:40:03 crc kubenswrapper[4813]: I0317 09:40:03.252577 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562340-cw8rp" podStartSLOduration=1.6390536610000002 podStartE2EDuration="3.25255609s" podCreationTimestamp="2026-03-17 09:40:00 +0000 UTC" firstStartedPulling="2026-03-17 09:40:01.065685439 +0000 UTC m=+1823.166488948" lastFinishedPulling="2026-03-17 09:40:02.679187838 +0000 UTC m=+1824.779991377" observedRunningTime="2026-03-17 09:40:03.238778384 +0000 UTC m=+1825.339581893" watchObservedRunningTime="2026-03-17 09:40:03.25255609 +0000 UTC m=+1825.353359609" Mar 17 09:40:04 crc kubenswrapper[4813]: I0317 09:40:04.237372 4813 generic.go:334] "Generic (PLEG): container finished" podID="8b5d72c0-ddd6-4c41-8845-7052c340a046" containerID="09b1e1e803fe358a61aa5bfed604dbac24f44be69505e18ae4d8b75ed810f988" exitCode=0 Mar 17 09:40:04 crc kubenswrapper[4813]: I0317 09:40:04.237454 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562340-cw8rp" event={"ID":"8b5d72c0-ddd6-4c41-8845-7052c340a046","Type":"ContainerDied","Data":"09b1e1e803fe358a61aa5bfed604dbac24f44be69505e18ae4d8b75ed810f988"} Mar 17 09:40:04 crc kubenswrapper[4813]: I0317 09:40:04.733542 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:40:04 crc kubenswrapper[4813]: E0317 09:40:04.734096 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:40:05 crc kubenswrapper[4813]: I0317 09:40:05.638996 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562340-cw8rp" Mar 17 09:40:05 crc kubenswrapper[4813]: I0317 09:40:05.701886 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4gqnh\" (UniqueName: \"kubernetes.io/projected/8b5d72c0-ddd6-4c41-8845-7052c340a046-kube-api-access-4gqnh\") pod \"8b5d72c0-ddd6-4c41-8845-7052c340a046\" (UID: \"8b5d72c0-ddd6-4c41-8845-7052c340a046\") " Mar 17 09:40:05 crc kubenswrapper[4813]: I0317 09:40:05.710213 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b5d72c0-ddd6-4c41-8845-7052c340a046-kube-api-access-4gqnh" (OuterVolumeSpecName: "kube-api-access-4gqnh") pod "8b5d72c0-ddd6-4c41-8845-7052c340a046" (UID: "8b5d72c0-ddd6-4c41-8845-7052c340a046"). InnerVolumeSpecName "kube-api-access-4gqnh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:40:05 crc kubenswrapper[4813]: I0317 09:40:05.804681 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4gqnh\" (UniqueName: \"kubernetes.io/projected/8b5d72c0-ddd6-4c41-8845-7052c340a046-kube-api-access-4gqnh\") on node \"crc\" DevicePath \"\"" Mar 17 09:40:06 crc kubenswrapper[4813]: I0317 09:40:06.046738 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-8w5d7"] Mar 17 09:40:06 crc kubenswrapper[4813]: I0317 09:40:06.058472 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-s6zhg"] Mar 17 09:40:06 crc kubenswrapper[4813]: I0317 09:40:06.067510 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-2d07-account-create-update-4j8f9"] Mar 17 09:40:06 crc kubenswrapper[4813]: I0317 09:40:06.085895 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-8w5d7"] Mar 17 09:40:06 crc kubenswrapper[4813]: I0317 09:40:06.094455 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-2d07-account-create-update-4j8f9"] Mar 17 09:40:06 crc kubenswrapper[4813]: I0317 09:40:06.104939 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-s6zhg"] Mar 17 09:40:06 crc kubenswrapper[4813]: I0317 09:40:06.264748 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562340-cw8rp" event={"ID":"8b5d72c0-ddd6-4c41-8845-7052c340a046","Type":"ContainerDied","Data":"93e5ab76653f205dee577192a71d2fbce73d9e52923d5d89bef31410ef717238"} Mar 17 09:40:06 crc kubenswrapper[4813]: I0317 09:40:06.264801 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="93e5ab76653f205dee577192a71d2fbce73d9e52923d5d89bef31410ef717238" Mar 17 09:40:06 crc kubenswrapper[4813]: I0317 09:40:06.264871 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562340-cw8rp" Mar 17 09:40:06 crc kubenswrapper[4813]: I0317 09:40:06.321938 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562334-bs752"] Mar 17 09:40:06 crc kubenswrapper[4813]: I0317 09:40:06.335454 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562334-bs752"] Mar 17 09:40:06 crc kubenswrapper[4813]: I0317 09:40:06.743200 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cbf9e10-aa11-4c2e-9930-b4f1f451bc65" path="/var/lib/kubelet/pods/0cbf9e10-aa11-4c2e-9930-b4f1f451bc65/volumes" Mar 17 09:40:06 crc kubenswrapper[4813]: I0317 09:40:06.765197 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d363247c-cdbf-49cc-b332-e87f849d9fee" path="/var/lib/kubelet/pods/d363247c-cdbf-49cc-b332-e87f849d9fee/volumes" Mar 17 09:40:06 crc kubenswrapper[4813]: I0317 09:40:06.766357 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7b2d571-5a93-4520-8bab-324e4d27bdc9" path="/var/lib/kubelet/pods/d7b2d571-5a93-4520-8bab-324e4d27bdc9/volumes" Mar 17 09:40:06 crc kubenswrapper[4813]: I0317 09:40:06.767414 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4b19fd6-39e5-49da-811a-81b0ce1f0c29" path="/var/lib/kubelet/pods/e4b19fd6-39e5-49da-811a-81b0ce1f0c29/volumes" Mar 17 09:40:16 crc kubenswrapper[4813]: I0317 09:40:16.375479 4813 generic.go:334] "Generic (PLEG): container finished" podID="36536921-03a1-45dd-93b3-1d06e9c3adca" containerID="fe26fb49a47dd7bdf5785d4e92a4e5c9c609ee0f77b9eafd13e7462bde10f26b" exitCode=0 Mar 17 09:40:16 crc kubenswrapper[4813]: I0317 09:40:16.375568 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46np8" event={"ID":"36536921-03a1-45dd-93b3-1d06e9c3adca","Type":"ContainerDied","Data":"fe26fb49a47dd7bdf5785d4e92a4e5c9c609ee0f77b9eafd13e7462bde10f26b"} Mar 17 09:40:16 crc kubenswrapper[4813]: I0317 09:40:16.731854 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:40:16 crc kubenswrapper[4813]: E0317 09:40:16.732407 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:40:17 crc kubenswrapper[4813]: I0317 09:40:17.857523 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46np8" Mar 17 09:40:17 crc kubenswrapper[4813]: I0317 09:40:17.968944 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/36536921-03a1-45dd-93b3-1d06e9c3adca-ssh-key-openstack-edpm-ipam\") pod \"36536921-03a1-45dd-93b3-1d06e9c3adca\" (UID: \"36536921-03a1-45dd-93b3-1d06e9c3adca\") " Mar 17 09:40:17 crc kubenswrapper[4813]: I0317 09:40:17.969046 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36536921-03a1-45dd-93b3-1d06e9c3adca-inventory\") pod \"36536921-03a1-45dd-93b3-1d06e9c3adca\" (UID: \"36536921-03a1-45dd-93b3-1d06e9c3adca\") " Mar 17 09:40:17 crc kubenswrapper[4813]: I0317 09:40:17.969104 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmdk7\" (UniqueName: \"kubernetes.io/projected/36536921-03a1-45dd-93b3-1d06e9c3adca-kube-api-access-qmdk7\") pod \"36536921-03a1-45dd-93b3-1d06e9c3adca\" (UID: \"36536921-03a1-45dd-93b3-1d06e9c3adca\") " Mar 17 09:40:17 crc kubenswrapper[4813]: I0317 09:40:17.976111 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36536921-03a1-45dd-93b3-1d06e9c3adca-kube-api-access-qmdk7" (OuterVolumeSpecName: "kube-api-access-qmdk7") pod "36536921-03a1-45dd-93b3-1d06e9c3adca" (UID: "36536921-03a1-45dd-93b3-1d06e9c3adca"). InnerVolumeSpecName "kube-api-access-qmdk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:40:18 crc kubenswrapper[4813]: E0317 09:40:18.020169 4813 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/36536921-03a1-45dd-93b3-1d06e9c3adca-inventory podName:36536921-03a1-45dd-93b3-1d06e9c3adca nodeName:}" failed. No retries permitted until 2026-03-17 09:40:18.520142605 +0000 UTC m=+1840.620946104 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "inventory" (UniqueName: "kubernetes.io/secret/36536921-03a1-45dd-93b3-1d06e9c3adca-inventory") pod "36536921-03a1-45dd-93b3-1d06e9c3adca" (UID: "36536921-03a1-45dd-93b3-1d06e9c3adca") : error deleting /var/lib/kubelet/pods/36536921-03a1-45dd-93b3-1d06e9c3adca/volume-subpaths: remove /var/lib/kubelet/pods/36536921-03a1-45dd-93b3-1d06e9c3adca/volume-subpaths: no such file or directory Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.023903 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36536921-03a1-45dd-93b3-1d06e9c3adca-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "36536921-03a1-45dd-93b3-1d06e9c3adca" (UID: "36536921-03a1-45dd-93b3-1d06e9c3adca"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.074219 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmdk7\" (UniqueName: \"kubernetes.io/projected/36536921-03a1-45dd-93b3-1d06e9c3adca-kube-api-access-qmdk7\") on node \"crc\" DevicePath \"\"" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.074254 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/36536921-03a1-45dd-93b3-1d06e9c3adca-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.399503 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46np8" event={"ID":"36536921-03a1-45dd-93b3-1d06e9c3adca","Type":"ContainerDied","Data":"35dda0fcf8586e0e107cce41b0e526075d81575b40b15b61dd5532b56201e720"} Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.399545 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35dda0fcf8586e0e107cce41b0e526075d81575b40b15b61dd5532b56201e720" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.399956 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/download-cache-edpm-deployment-openstack-edpm-ipam-46np8" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.501030 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4kzll"] Mar 17 09:40:18 crc kubenswrapper[4813]: E0317 09:40:18.501643 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b5d72c0-ddd6-4c41-8845-7052c340a046" containerName="oc" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.501670 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b5d72c0-ddd6-4c41-8845-7052c340a046" containerName="oc" Mar 17 09:40:18 crc kubenswrapper[4813]: E0317 09:40:18.501711 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36536921-03a1-45dd-93b3-1d06e9c3adca" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.501725 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="36536921-03a1-45dd-93b3-1d06e9c3adca" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.502029 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="36536921-03a1-45dd-93b3-1d06e9c3adca" containerName="download-cache-edpm-deployment-openstack-edpm-ipam" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.502067 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b5d72c0-ddd6-4c41-8845-7052c340a046" containerName="oc" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.503277 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4kzll" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.512411 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4kzll"] Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.585068 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36536921-03a1-45dd-93b3-1d06e9c3adca-inventory\") pod \"36536921-03a1-45dd-93b3-1d06e9c3adca\" (UID: \"36536921-03a1-45dd-93b3-1d06e9c3adca\") " Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.585473 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bnfdq\" (UniqueName: \"kubernetes.io/projected/c7908d9c-b027-4b03-8879-37968eea2f28-kube-api-access-bnfdq\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4kzll\" (UID: \"c7908d9c-b027-4b03-8879-37968eea2f28\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4kzll" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.585749 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c7908d9c-b027-4b03-8879-37968eea2f28-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4kzll\" (UID: \"c7908d9c-b027-4b03-8879-37968eea2f28\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4kzll" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.585799 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c7908d9c-b027-4b03-8879-37968eea2f28-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4kzll\" (UID: \"c7908d9c-b027-4b03-8879-37968eea2f28\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4kzll" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.587697 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/36536921-03a1-45dd-93b3-1d06e9c3adca-inventory" (OuterVolumeSpecName: "inventory") pod "36536921-03a1-45dd-93b3-1d06e9c3adca" (UID: "36536921-03a1-45dd-93b3-1d06e9c3adca"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.687621 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bnfdq\" (UniqueName: \"kubernetes.io/projected/c7908d9c-b027-4b03-8879-37968eea2f28-kube-api-access-bnfdq\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4kzll\" (UID: \"c7908d9c-b027-4b03-8879-37968eea2f28\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4kzll" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.687723 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c7908d9c-b027-4b03-8879-37968eea2f28-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4kzll\" (UID: \"c7908d9c-b027-4b03-8879-37968eea2f28\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4kzll" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.687767 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c7908d9c-b027-4b03-8879-37968eea2f28-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4kzll\" (UID: \"c7908d9c-b027-4b03-8879-37968eea2f28\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4kzll" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.687849 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/36536921-03a1-45dd-93b3-1d06e9c3adca-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.690807 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c7908d9c-b027-4b03-8879-37968eea2f28-ssh-key-openstack-edpm-ipam\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4kzll\" (UID: \"c7908d9c-b027-4b03-8879-37968eea2f28\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4kzll" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.692523 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c7908d9c-b027-4b03-8879-37968eea2f28-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4kzll\" (UID: \"c7908d9c-b027-4b03-8879-37968eea2f28\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4kzll" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.713743 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bnfdq\" (UniqueName: \"kubernetes.io/projected/c7908d9c-b027-4b03-8879-37968eea2f28-kube-api-access-bnfdq\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4kzll\" (UID: \"c7908d9c-b027-4b03-8879-37968eea2f28\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4kzll" Mar 17 09:40:18 crc kubenswrapper[4813]: I0317 09:40:18.867855 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4kzll" Mar 17 09:40:19 crc kubenswrapper[4813]: I0317 09:40:19.468916 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4kzll"] Mar 17 09:40:20 crc kubenswrapper[4813]: I0317 09:40:20.437067 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4kzll" event={"ID":"c7908d9c-b027-4b03-8879-37968eea2f28","Type":"ContainerStarted","Data":"24d18a28fab29c3863b1f9037c2fa7ab4872245b8e1c7fd8464619a97baf420f"} Mar 17 09:40:20 crc kubenswrapper[4813]: I0317 09:40:20.437377 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4kzll" event={"ID":"c7908d9c-b027-4b03-8879-37968eea2f28","Type":"ContainerStarted","Data":"4cca79e8e57d9310aa3236f450e63cbeee975c97f633c5f18342058875e5b448"} Mar 17 09:40:20 crc kubenswrapper[4813]: I0317 09:40:20.460786 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4kzll" podStartSLOduration=1.861892028 podStartE2EDuration="2.460757527s" podCreationTimestamp="2026-03-17 09:40:18 +0000 UTC" firstStartedPulling="2026-03-17 09:40:19.474191903 +0000 UTC m=+1841.574995412" lastFinishedPulling="2026-03-17 09:40:20.073057392 +0000 UTC m=+1842.173860911" observedRunningTime="2026-03-17 09:40:20.452717334 +0000 UTC m=+1842.553520843" watchObservedRunningTime="2026-03-17 09:40:20.460757527 +0000 UTC m=+1842.561561056" Mar 17 09:40:26 crc kubenswrapper[4813]: I0317 09:40:26.073339 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-nrh2h"] Mar 17 09:40:26 crc kubenswrapper[4813]: I0317 09:40:26.085147 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-nrh2h"] Mar 17 09:40:26 crc kubenswrapper[4813]: I0317 09:40:26.748360 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8357c2d-7bfc-4296-8d39-7f64f573aa45" path="/var/lib/kubelet/pods/b8357c2d-7bfc-4296-8d39-7f64f573aa45/volumes" Mar 17 09:40:31 crc kubenswrapper[4813]: I0317 09:40:31.039283 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/root-account-create-update-drprf"] Mar 17 09:40:31 crc kubenswrapper[4813]: I0317 09:40:31.048293 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/root-account-create-update-drprf"] Mar 17 09:40:31 crc kubenswrapper[4813]: I0317 09:40:31.730741 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:40:31 crc kubenswrapper[4813]: E0317 09:40:31.732089 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:40:32 crc kubenswrapper[4813]: I0317 09:40:32.750554 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="075476d6-2221-4309-9e08-3d9d4f4bdfe1" path="/var/lib/kubelet/pods/075476d6-2221-4309-9e08-3d9d4f4bdfe1/volumes" Mar 17 09:40:43 crc kubenswrapper[4813]: I0317 09:40:43.731212 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:40:43 crc kubenswrapper[4813]: E0317 09:40:43.732388 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:40:44 crc kubenswrapper[4813]: I0317 09:40:44.056510 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-79mzp"] Mar 17 09:40:44 crc kubenswrapper[4813]: I0317 09:40:44.074062 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-8098-account-create-update-j4khm"] Mar 17 09:40:44 crc kubenswrapper[4813]: I0317 09:40:44.088777 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-79mzp"] Mar 17 09:40:44 crc kubenswrapper[4813]: I0317 09:40:44.100816 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-b8kql"] Mar 17 09:40:44 crc kubenswrapper[4813]: I0317 09:40:44.108222 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-8098-account-create-update-j4khm"] Mar 17 09:40:44 crc kubenswrapper[4813]: I0317 09:40:44.114583 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-b8kql"] Mar 17 09:40:44 crc kubenswrapper[4813]: I0317 09:40:44.121023 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-f43a-account-create-update-mxdk5"] Mar 17 09:40:44 crc kubenswrapper[4813]: I0317 09:40:44.127591 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-pn9ft"] Mar 17 09:40:44 crc kubenswrapper[4813]: I0317 09:40:44.133934 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-f43a-account-create-update-mxdk5"] Mar 17 09:40:44 crc kubenswrapper[4813]: I0317 09:40:44.140384 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-e948-account-create-update-prpwx"] Mar 17 09:40:44 crc kubenswrapper[4813]: I0317 09:40:44.146890 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-pn9ft"] Mar 17 09:40:44 crc kubenswrapper[4813]: I0317 09:40:44.152997 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-e948-account-create-update-prpwx"] Mar 17 09:40:44 crc kubenswrapper[4813]: I0317 09:40:44.747576 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="34bc865e-b877-44b4-a57d-d7f390f60414" path="/var/lib/kubelet/pods/34bc865e-b877-44b4-a57d-d7f390f60414/volumes" Mar 17 09:40:44 crc kubenswrapper[4813]: I0317 09:40:44.748882 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3666f515-3ed8-486a-a775-eea5c4463580" path="/var/lib/kubelet/pods/3666f515-3ed8-486a-a775-eea5c4463580/volumes" Mar 17 09:40:44 crc kubenswrapper[4813]: I0317 09:40:44.749569 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a047928-f15b-4c01-a4ef-b1c4b4d45cc9" path="/var/lib/kubelet/pods/5a047928-f15b-4c01-a4ef-b1c4b4d45cc9/volumes" Mar 17 09:40:44 crc kubenswrapper[4813]: I0317 09:40:44.750302 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6cb4aebe-be10-4574-8851-993dc42a3759" path="/var/lib/kubelet/pods/6cb4aebe-be10-4574-8851-993dc42a3759/volumes" Mar 17 09:40:44 crc kubenswrapper[4813]: I0317 09:40:44.751764 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a10880b9-b0fb-4510-a51a-c93bbb6aa563" path="/var/lib/kubelet/pods/a10880b9-b0fb-4510-a51a-c93bbb6aa563/volumes" Mar 17 09:40:44 crc kubenswrapper[4813]: I0317 09:40:44.752627 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae41df76-1944-4060-81a1-b28c7c839faf" path="/var/lib/kubelet/pods/ae41df76-1944-4060-81a1-b28c7c839faf/volumes" Mar 17 09:40:46 crc kubenswrapper[4813]: I0317 09:40:46.390991 4813 scope.go:117] "RemoveContainer" containerID="b29e9b4eda5417c83566c4aafd4a22ca559e43d575dcc2c9328f0e67d69d820c" Mar 17 09:40:46 crc kubenswrapper[4813]: I0317 09:40:46.419660 4813 scope.go:117] "RemoveContainer" containerID="18dad9114889c2dcbe23adb0769418a921cf69093eb881946fbd29b93b802eb0" Mar 17 09:40:46 crc kubenswrapper[4813]: I0317 09:40:46.462476 4813 scope.go:117] "RemoveContainer" containerID="bed2240cb56ec958df9556f000c568cf41aaf1e047d6713adee45fc9e40288b1" Mar 17 09:40:46 crc kubenswrapper[4813]: I0317 09:40:46.507748 4813 scope.go:117] "RemoveContainer" containerID="917eabd06202b2d1648e0f49f1227c65f4e00a2c5d4af77ea4e9b631ae4cfce5" Mar 17 09:40:46 crc kubenswrapper[4813]: I0317 09:40:46.577510 4813 scope.go:117] "RemoveContainer" containerID="3759fe17e8df5399c3f9362f787395f71f578aa9dd86d678642fc1e444352438" Mar 17 09:40:46 crc kubenswrapper[4813]: I0317 09:40:46.639275 4813 scope.go:117] "RemoveContainer" containerID="e4a14c386c4f92a231421b0f9a45151032bd59a580677087d5c334994723de64" Mar 17 09:40:46 crc kubenswrapper[4813]: I0317 09:40:46.696251 4813 scope.go:117] "RemoveContainer" containerID="26fce0a124538082a35190759f3217c42a5039ec4a698e529f898ce6480af7e7" Mar 17 09:40:46 crc kubenswrapper[4813]: I0317 09:40:46.748626 4813 scope.go:117] "RemoveContainer" containerID="c6452ad9a30c1975243728d5c7fab8eae68073e39d7f4fdf834c41edbbf86c44" Mar 17 09:40:46 crc kubenswrapper[4813]: I0317 09:40:46.776025 4813 scope.go:117] "RemoveContainer" containerID="cced6e2efb9888502f8215f475b217859c218afc2a981b90b782f833493ba942" Mar 17 09:40:46 crc kubenswrapper[4813]: I0317 09:40:46.808246 4813 scope.go:117] "RemoveContainer" containerID="87f5d41a069b50be3e09350e28a6e33d90b1d628153073fab9b9017b3d1e78af" Mar 17 09:40:46 crc kubenswrapper[4813]: I0317 09:40:46.832921 4813 scope.go:117] "RemoveContainer" containerID="99c90a851496fedf51f09757cf22a17306cec4664540c90f864f16dc379cec24" Mar 17 09:40:46 crc kubenswrapper[4813]: I0317 09:40:46.874022 4813 scope.go:117] "RemoveContainer" containerID="77f9b0dc665da47be2b72e324542837824014ca8377b7d1a66110ff7f2ba8f38" Mar 17 09:40:46 crc kubenswrapper[4813]: I0317 09:40:46.929721 4813 scope.go:117] "RemoveContainer" containerID="8d6b382a460b8fffdc35f8457f0856fa2144685a719d1c467f61d7e75a46919f" Mar 17 09:40:46 crc kubenswrapper[4813]: I0317 09:40:46.984860 4813 scope.go:117] "RemoveContainer" containerID="19feba5cde4b9c9e1abb5c6c7468776a0c4d71ef6ce95e9dc53af16aa0549546" Mar 17 09:40:47 crc kubenswrapper[4813]: I0317 09:40:47.015219 4813 scope.go:117] "RemoveContainer" containerID="7de276f53292a53bf05b1f8194df75b83e957f1cb6a3b52b9cd3f27bb0f7c3d3" Mar 17 09:40:49 crc kubenswrapper[4813]: I0317 09:40:49.031047 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-xlbs6"] Mar 17 09:40:49 crc kubenswrapper[4813]: I0317 09:40:49.042207 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-xlbs6"] Mar 17 09:40:50 crc kubenswrapper[4813]: I0317 09:40:50.746119 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f531bb05-ff10-4a9b-a8f4-58dc28a8007a" path="/var/lib/kubelet/pods/f531bb05-ff10-4a9b-a8f4-58dc28a8007a/volumes" Mar 17 09:40:57 crc kubenswrapper[4813]: I0317 09:40:57.829585 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:40:57 crc kubenswrapper[4813]: E0317 09:40:57.830279 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:41:09 crc kubenswrapper[4813]: I0317 09:41:09.733197 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:41:09 crc kubenswrapper[4813]: E0317 09:41:09.734261 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:41:20 crc kubenswrapper[4813]: I0317 09:41:20.064146 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-jtbdz"] Mar 17 09:41:20 crc kubenswrapper[4813]: I0317 09:41:20.083626 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-jtbdz"] Mar 17 09:41:20 crc kubenswrapper[4813]: I0317 09:41:20.747358 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="765fce5e-7101-42a5-8ad1-172c967fd2c3" path="/var/lib/kubelet/pods/765fce5e-7101-42a5-8ad1-172c967fd2c3/volumes" Mar 17 09:41:24 crc kubenswrapper[4813]: I0317 09:41:24.731856 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:41:24 crc kubenswrapper[4813]: E0317 09:41:24.732550 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:41:25 crc kubenswrapper[4813]: I0317 09:41:25.043418 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-hkmc6"] Mar 17 09:41:25 crc kubenswrapper[4813]: I0317 09:41:25.059618 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-hkmc6"] Mar 17 09:41:26 crc kubenswrapper[4813]: I0317 09:41:26.749756 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c613a9b-36b2-437d-9c11-252d7d8ce910" path="/var/lib/kubelet/pods/9c613a9b-36b2-437d-9c11-252d7d8ce910/volumes" Mar 17 09:41:27 crc kubenswrapper[4813]: I0317 09:41:27.199396 4813 generic.go:334] "Generic (PLEG): container finished" podID="c7908d9c-b027-4b03-8879-37968eea2f28" containerID="24d18a28fab29c3863b1f9037c2fa7ab4872245b8e1c7fd8464619a97baf420f" exitCode=0 Mar 17 09:41:27 crc kubenswrapper[4813]: I0317 09:41:27.199497 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4kzll" event={"ID":"c7908d9c-b027-4b03-8879-37968eea2f28","Type":"ContainerDied","Data":"24d18a28fab29c3863b1f9037c2fa7ab4872245b8e1c7fd8464619a97baf420f"} Mar 17 09:41:28 crc kubenswrapper[4813]: I0317 09:41:28.058761 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-fdxt8"] Mar 17 09:41:28 crc kubenswrapper[4813]: I0317 09:41:28.069767 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-fdxt8"] Mar 17 09:41:28 crc kubenswrapper[4813]: I0317 09:41:28.647387 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4kzll" Mar 17 09:41:28 crc kubenswrapper[4813]: I0317 09:41:28.741233 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00b72575-1230-444b-9c9b-6c2e2cc921f0" path="/var/lib/kubelet/pods/00b72575-1230-444b-9c9b-6c2e2cc921f0/volumes" Mar 17 09:41:28 crc kubenswrapper[4813]: I0317 09:41:28.755123 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c7908d9c-b027-4b03-8879-37968eea2f28-ssh-key-openstack-edpm-ipam\") pod \"c7908d9c-b027-4b03-8879-37968eea2f28\" (UID: \"c7908d9c-b027-4b03-8879-37968eea2f28\") " Mar 17 09:41:28 crc kubenswrapper[4813]: I0317 09:41:28.755233 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c7908d9c-b027-4b03-8879-37968eea2f28-inventory\") pod \"c7908d9c-b027-4b03-8879-37968eea2f28\" (UID: \"c7908d9c-b027-4b03-8879-37968eea2f28\") " Mar 17 09:41:28 crc kubenswrapper[4813]: I0317 09:41:28.755289 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bnfdq\" (UniqueName: \"kubernetes.io/projected/c7908d9c-b027-4b03-8879-37968eea2f28-kube-api-access-bnfdq\") pod \"c7908d9c-b027-4b03-8879-37968eea2f28\" (UID: \"c7908d9c-b027-4b03-8879-37968eea2f28\") " Mar 17 09:41:28 crc kubenswrapper[4813]: I0317 09:41:28.764075 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c7908d9c-b027-4b03-8879-37968eea2f28-kube-api-access-bnfdq" (OuterVolumeSpecName: "kube-api-access-bnfdq") pod "c7908d9c-b027-4b03-8879-37968eea2f28" (UID: "c7908d9c-b027-4b03-8879-37968eea2f28"). InnerVolumeSpecName "kube-api-access-bnfdq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:41:28 crc kubenswrapper[4813]: I0317 09:41:28.792179 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7908d9c-b027-4b03-8879-37968eea2f28-inventory" (OuterVolumeSpecName: "inventory") pod "c7908d9c-b027-4b03-8879-37968eea2f28" (UID: "c7908d9c-b027-4b03-8879-37968eea2f28"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:41:28 crc kubenswrapper[4813]: I0317 09:41:28.803427 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c7908d9c-b027-4b03-8879-37968eea2f28-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "c7908d9c-b027-4b03-8879-37968eea2f28" (UID: "c7908d9c-b027-4b03-8879-37968eea2f28"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:41:28 crc kubenswrapper[4813]: I0317 09:41:28.857857 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/c7908d9c-b027-4b03-8879-37968eea2f28-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 09:41:28 crc kubenswrapper[4813]: I0317 09:41:28.857889 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c7908d9c-b027-4b03-8879-37968eea2f28-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 09:41:28 crc kubenswrapper[4813]: I0317 09:41:28.857898 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bnfdq\" (UniqueName: \"kubernetes.io/projected/c7908d9c-b027-4b03-8879-37968eea2f28-kube-api-access-bnfdq\") on node \"crc\" DevicePath \"\"" Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.226716 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4kzll" event={"ID":"c7908d9c-b027-4b03-8879-37968eea2f28","Type":"ContainerDied","Data":"4cca79e8e57d9310aa3236f450e63cbeee975c97f633c5f18342058875e5b448"} Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.226779 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4cca79e8e57d9310aa3236f450e63cbeee975c97f633c5f18342058875e5b448" Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.226806 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4kzll" Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.477070 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm"] Mar 17 09:41:29 crc kubenswrapper[4813]: E0317 09:41:29.477663 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c7908d9c-b027-4b03-8879-37968eea2f28" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.477692 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c7908d9c-b027-4b03-8879-37968eea2f28" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.478009 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c7908d9c-b027-4b03-8879-37968eea2f28" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.478880 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm" Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.482763 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hj9h4" Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.482838 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.482932 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.484917 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.492368 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm"] Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.578729 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c2d5562-76da-433b-9840-59385eeb872c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm\" (UID: \"0c2d5562-76da-433b-9840-59385eeb872c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm" Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.578829 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h48s7\" (UniqueName: \"kubernetes.io/projected/0c2d5562-76da-433b-9840-59385eeb872c-kube-api-access-h48s7\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm\" (UID: \"0c2d5562-76da-433b-9840-59385eeb872c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm" Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.579021 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0c2d5562-76da-433b-9840-59385eeb872c-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm\" (UID: \"0c2d5562-76da-433b-9840-59385eeb872c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm" Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.681347 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h48s7\" (UniqueName: \"kubernetes.io/projected/0c2d5562-76da-433b-9840-59385eeb872c-kube-api-access-h48s7\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm\" (UID: \"0c2d5562-76da-433b-9840-59385eeb872c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm" Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.682810 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0c2d5562-76da-433b-9840-59385eeb872c-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm\" (UID: \"0c2d5562-76da-433b-9840-59385eeb872c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm" Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.684778 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c2d5562-76da-433b-9840-59385eeb872c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm\" (UID: \"0c2d5562-76da-433b-9840-59385eeb872c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm" Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.690511 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0c2d5562-76da-433b-9840-59385eeb872c-ssh-key-openstack-edpm-ipam\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm\" (UID: \"0c2d5562-76da-433b-9840-59385eeb872c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm" Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.704727 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h48s7\" (UniqueName: \"kubernetes.io/projected/0c2d5562-76da-433b-9840-59385eeb872c-kube-api-access-h48s7\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm\" (UID: \"0c2d5562-76da-433b-9840-59385eeb872c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm" Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.708221 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c2d5562-76da-433b-9840-59385eeb872c-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm\" (UID: \"0c2d5562-76da-433b-9840-59385eeb872c\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm" Mar 17 09:41:29 crc kubenswrapper[4813]: I0317 09:41:29.810987 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm" Mar 17 09:41:30 crc kubenswrapper[4813]: I0317 09:41:30.159568 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm"] Mar 17 09:41:30 crc kubenswrapper[4813]: I0317 09:41:30.160416 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 09:41:30 crc kubenswrapper[4813]: I0317 09:41:30.237847 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm" event={"ID":"0c2d5562-76da-433b-9840-59385eeb872c","Type":"ContainerStarted","Data":"de0035eccac023e172c79fea92fda13bd0d12fa708d2e9632c39bb9896ac4254"} Mar 17 09:41:31 crc kubenswrapper[4813]: I0317 09:41:31.248190 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm" event={"ID":"0c2d5562-76da-433b-9840-59385eeb872c","Type":"ContainerStarted","Data":"a0e588a098a9d978dcf4a34cdd79b42427adb2b7520d3d0ee52106e7daf71edd"} Mar 17 09:41:35 crc kubenswrapper[4813]: I0317 09:41:35.289260 4813 generic.go:334] "Generic (PLEG): container finished" podID="0c2d5562-76da-433b-9840-59385eeb872c" containerID="a0e588a098a9d978dcf4a34cdd79b42427adb2b7520d3d0ee52106e7daf71edd" exitCode=0 Mar 17 09:41:35 crc kubenswrapper[4813]: I0317 09:41:35.289352 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm" event={"ID":"0c2d5562-76da-433b-9840-59385eeb872c","Type":"ContainerDied","Data":"a0e588a098a9d978dcf4a34cdd79b42427adb2b7520d3d0ee52106e7daf71edd"} Mar 17 09:41:36 crc kubenswrapper[4813]: I0317 09:41:36.693841 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm" Mar 17 09:41:36 crc kubenswrapper[4813]: I0317 09:41:36.734382 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:41:36 crc kubenswrapper[4813]: E0317 09:41:36.734887 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:41:36 crc kubenswrapper[4813]: I0317 09:41:36.764860 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c2d5562-76da-433b-9840-59385eeb872c-inventory\") pod \"0c2d5562-76da-433b-9840-59385eeb872c\" (UID: \"0c2d5562-76da-433b-9840-59385eeb872c\") " Mar 17 09:41:36 crc kubenswrapper[4813]: I0317 09:41:36.765076 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h48s7\" (UniqueName: \"kubernetes.io/projected/0c2d5562-76da-433b-9840-59385eeb872c-kube-api-access-h48s7\") pod \"0c2d5562-76da-433b-9840-59385eeb872c\" (UID: \"0c2d5562-76da-433b-9840-59385eeb872c\") " Mar 17 09:41:36 crc kubenswrapper[4813]: I0317 09:41:36.765157 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0c2d5562-76da-433b-9840-59385eeb872c-ssh-key-openstack-edpm-ipam\") pod \"0c2d5562-76da-433b-9840-59385eeb872c\" (UID: \"0c2d5562-76da-433b-9840-59385eeb872c\") " Mar 17 09:41:36 crc kubenswrapper[4813]: I0317 09:41:36.772440 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c2d5562-76da-433b-9840-59385eeb872c-kube-api-access-h48s7" (OuterVolumeSpecName: "kube-api-access-h48s7") pod "0c2d5562-76da-433b-9840-59385eeb872c" (UID: "0c2d5562-76da-433b-9840-59385eeb872c"). InnerVolumeSpecName "kube-api-access-h48s7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:41:36 crc kubenswrapper[4813]: I0317 09:41:36.802054 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c2d5562-76da-433b-9840-59385eeb872c-inventory" (OuterVolumeSpecName: "inventory") pod "0c2d5562-76da-433b-9840-59385eeb872c" (UID: "0c2d5562-76da-433b-9840-59385eeb872c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:41:36 crc kubenswrapper[4813]: I0317 09:41:36.811888 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c2d5562-76da-433b-9840-59385eeb872c-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "0c2d5562-76da-433b-9840-59385eeb872c" (UID: "0c2d5562-76da-433b-9840-59385eeb872c"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:41:36 crc kubenswrapper[4813]: I0317 09:41:36.867883 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0c2d5562-76da-433b-9840-59385eeb872c-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 09:41:36 crc kubenswrapper[4813]: I0317 09:41:36.867928 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h48s7\" (UniqueName: \"kubernetes.io/projected/0c2d5562-76da-433b-9840-59385eeb872c-kube-api-access-h48s7\") on node \"crc\" DevicePath \"\"" Mar 17 09:41:36 crc kubenswrapper[4813]: I0317 09:41:36.867943 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0c2d5562-76da-433b-9840-59385eeb872c-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.317064 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm" event={"ID":"0c2d5562-76da-433b-9840-59385eeb872c","Type":"ContainerDied","Data":"de0035eccac023e172c79fea92fda13bd0d12fa708d2e9632c39bb9896ac4254"} Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.317130 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="de0035eccac023e172c79fea92fda13bd0d12fa708d2e9632c39bb9896ac4254" Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.317166 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm" Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.402960 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-64lbt"] Mar 17 09:41:37 crc kubenswrapper[4813]: E0317 09:41:37.403658 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c2d5562-76da-433b-9840-59385eeb872c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.403687 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c2d5562-76da-433b-9840-59385eeb872c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.404029 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c2d5562-76da-433b-9840-59385eeb872c" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.405009 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-64lbt" Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.408995 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.409050 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hj9h4" Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.409204 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.410706 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.417475 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-64lbt"] Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.480844 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/63ad7386-248c-4dc8-a217-fb2cf3d4ef82-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-64lbt\" (UID: \"63ad7386-248c-4dc8-a217-fb2cf3d4ef82\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-64lbt" Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.480924 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/63ad7386-248c-4dc8-a217-fb2cf3d4ef82-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-64lbt\" (UID: \"63ad7386-248c-4dc8-a217-fb2cf3d4ef82\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-64lbt" Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.481012 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rnpp\" (UniqueName: \"kubernetes.io/projected/63ad7386-248c-4dc8-a217-fb2cf3d4ef82-kube-api-access-5rnpp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-64lbt\" (UID: \"63ad7386-248c-4dc8-a217-fb2cf3d4ef82\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-64lbt" Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.582111 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/63ad7386-248c-4dc8-a217-fb2cf3d4ef82-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-64lbt\" (UID: \"63ad7386-248c-4dc8-a217-fb2cf3d4ef82\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-64lbt" Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.582169 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/63ad7386-248c-4dc8-a217-fb2cf3d4ef82-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-64lbt\" (UID: \"63ad7386-248c-4dc8-a217-fb2cf3d4ef82\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-64lbt" Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.582228 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rnpp\" (UniqueName: \"kubernetes.io/projected/63ad7386-248c-4dc8-a217-fb2cf3d4ef82-kube-api-access-5rnpp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-64lbt\" (UID: \"63ad7386-248c-4dc8-a217-fb2cf3d4ef82\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-64lbt" Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.586583 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/63ad7386-248c-4dc8-a217-fb2cf3d4ef82-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-64lbt\" (UID: \"63ad7386-248c-4dc8-a217-fb2cf3d4ef82\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-64lbt" Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.588868 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/63ad7386-248c-4dc8-a217-fb2cf3d4ef82-ssh-key-openstack-edpm-ipam\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-64lbt\" (UID: \"63ad7386-248c-4dc8-a217-fb2cf3d4ef82\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-64lbt" Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.598475 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rnpp\" (UniqueName: \"kubernetes.io/projected/63ad7386-248c-4dc8-a217-fb2cf3d4ef82-kube-api-access-5rnpp\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-64lbt\" (UID: \"63ad7386-248c-4dc8-a217-fb2cf3d4ef82\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-64lbt" Mar 17 09:41:37 crc kubenswrapper[4813]: I0317 09:41:37.721703 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-64lbt" Mar 17 09:41:38 crc kubenswrapper[4813]: I0317 09:41:38.390519 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-64lbt"] Mar 17 09:41:38 crc kubenswrapper[4813]: W0317 09:41:38.393879 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod63ad7386_248c_4dc8_a217_fb2cf3d4ef82.slice/crio-7d0011186339754aaedff52451f938a1cb9db2ddf81db1dea5f5756216a045e4 WatchSource:0}: Error finding container 7d0011186339754aaedff52451f938a1cb9db2ddf81db1dea5f5756216a045e4: Status 404 returned error can't find the container with id 7d0011186339754aaedff52451f938a1cb9db2ddf81db1dea5f5756216a045e4 Mar 17 09:41:38 crc kubenswrapper[4813]: I0317 09:41:38.814377 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 09:41:39 crc kubenswrapper[4813]: I0317 09:41:39.342771 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-64lbt" event={"ID":"63ad7386-248c-4dc8-a217-fb2cf3d4ef82","Type":"ContainerStarted","Data":"74f20b605ba23ca7e4078d252d3d9162a851d1b865f0427cee85376a5beac585"} Mar 17 09:41:39 crc kubenswrapper[4813]: I0317 09:41:39.343136 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-64lbt" event={"ID":"63ad7386-248c-4dc8-a217-fb2cf3d4ef82","Type":"ContainerStarted","Data":"7d0011186339754aaedff52451f938a1cb9db2ddf81db1dea5f5756216a045e4"} Mar 17 09:41:39 crc kubenswrapper[4813]: I0317 09:41:39.366566 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-64lbt" podStartSLOduration=1.950397511 podStartE2EDuration="2.366538147s" podCreationTimestamp="2026-03-17 09:41:37 +0000 UTC" firstStartedPulling="2026-03-17 09:41:38.396020695 +0000 UTC m=+1920.496824194" lastFinishedPulling="2026-03-17 09:41:38.812161331 +0000 UTC m=+1920.912964830" observedRunningTime="2026-03-17 09:41:39.360513458 +0000 UTC m=+1921.461316997" watchObservedRunningTime="2026-03-17 09:41:39.366538147 +0000 UTC m=+1921.467341656" Mar 17 09:41:45 crc kubenswrapper[4813]: I0317 09:41:45.045441 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-mxghx"] Mar 17 09:41:45 crc kubenswrapper[4813]: I0317 09:41:45.055889 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-27g9l"] Mar 17 09:41:45 crc kubenswrapper[4813]: I0317 09:41:45.065685 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-27g9l"] Mar 17 09:41:45 crc kubenswrapper[4813]: I0317 09:41:45.076395 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-mxghx"] Mar 17 09:41:46 crc kubenswrapper[4813]: I0317 09:41:46.752896 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41809604-4a36-4453-b381-66b4c5cf1c43" path="/var/lib/kubelet/pods/41809604-4a36-4453-b381-66b4c5cf1c43/volumes" Mar 17 09:41:46 crc kubenswrapper[4813]: I0317 09:41:46.755074 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c6cd309-922e-4be6-b2d9-12d2aca54b78" path="/var/lib/kubelet/pods/5c6cd309-922e-4be6-b2d9-12d2aca54b78/volumes" Mar 17 09:41:47 crc kubenswrapper[4813]: I0317 09:41:47.329189 4813 scope.go:117] "RemoveContainer" containerID="f95998cbf5bb06f0793dbfffb2b455c0da8571e30177362a52d817b3aefc91d8" Mar 17 09:41:47 crc kubenswrapper[4813]: I0317 09:41:47.380198 4813 scope.go:117] "RemoveContainer" containerID="fc487a06c5503d88b64ffd200cae2436b83e8922ce174dff8d881b7437ea83c0" Mar 17 09:41:47 crc kubenswrapper[4813]: I0317 09:41:47.433658 4813 scope.go:117] "RemoveContainer" containerID="c0bcfc07898a4177c7b334fd23c56e6fbd4fa2af6b58ab9ba3a9942136ae713b" Mar 17 09:41:47 crc kubenswrapper[4813]: I0317 09:41:47.506670 4813 scope.go:117] "RemoveContainer" containerID="3933cb26fc972a9b167b78a49896693ed6615d62ac3e9962d795dde203e9cd3f" Mar 17 09:41:47 crc kubenswrapper[4813]: I0317 09:41:47.548585 4813 scope.go:117] "RemoveContainer" containerID="7d2252f8de6b1e8bafbbf5cf0d28de406e54ae318f14627bf88dbb57fc4ef071" Mar 17 09:41:47 crc kubenswrapper[4813]: I0317 09:41:47.588339 4813 scope.go:117] "RemoveContainer" containerID="3e5b3f1123b7a79595271d2af94d410f821a75be59a0cf9ea6e77c1f02f6edee" Mar 17 09:41:47 crc kubenswrapper[4813]: I0317 09:41:47.629286 4813 scope.go:117] "RemoveContainer" containerID="bd4a1f290d52c558f9005e068c1bc067598630436ad751660d444966ab779596" Mar 17 09:41:47 crc kubenswrapper[4813]: I0317 09:41:47.665051 4813 scope.go:117] "RemoveContainer" containerID="2e13e45a5cbf8e2e08800c4276301a0d39702765d34df2bcb61d04f52b11d8a9" Mar 17 09:41:50 crc kubenswrapper[4813]: I0317 09:41:50.732260 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:41:51 crc kubenswrapper[4813]: I0317 09:41:51.481462 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerStarted","Data":"886e3a29e582cbf7efd8021b4dad6057b4824ee4ecc8a4ac9ae12df4687c6449"} Mar 17 09:42:00 crc kubenswrapper[4813]: I0317 09:42:00.169934 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562342-bq4l7"] Mar 17 09:42:00 crc kubenswrapper[4813]: I0317 09:42:00.172060 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562342-bq4l7" Mar 17 09:42:00 crc kubenswrapper[4813]: I0317 09:42:00.174708 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:42:00 crc kubenswrapper[4813]: I0317 09:42:00.175162 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:42:00 crc kubenswrapper[4813]: I0317 09:42:00.175456 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:42:00 crc kubenswrapper[4813]: I0317 09:42:00.179634 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562342-bq4l7"] Mar 17 09:42:00 crc kubenswrapper[4813]: I0317 09:42:00.314109 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgtcw\" (UniqueName: \"kubernetes.io/projected/0b2f2eb2-09cd-422d-80a1-630d2811c390-kube-api-access-bgtcw\") pod \"auto-csr-approver-29562342-bq4l7\" (UID: \"0b2f2eb2-09cd-422d-80a1-630d2811c390\") " pod="openshift-infra/auto-csr-approver-29562342-bq4l7" Mar 17 09:42:00 crc kubenswrapper[4813]: I0317 09:42:00.416685 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgtcw\" (UniqueName: \"kubernetes.io/projected/0b2f2eb2-09cd-422d-80a1-630d2811c390-kube-api-access-bgtcw\") pod \"auto-csr-approver-29562342-bq4l7\" (UID: \"0b2f2eb2-09cd-422d-80a1-630d2811c390\") " pod="openshift-infra/auto-csr-approver-29562342-bq4l7" Mar 17 09:42:00 crc kubenswrapper[4813]: I0317 09:42:00.440258 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgtcw\" (UniqueName: \"kubernetes.io/projected/0b2f2eb2-09cd-422d-80a1-630d2811c390-kube-api-access-bgtcw\") pod \"auto-csr-approver-29562342-bq4l7\" (UID: \"0b2f2eb2-09cd-422d-80a1-630d2811c390\") " pod="openshift-infra/auto-csr-approver-29562342-bq4l7" Mar 17 09:42:00 crc kubenswrapper[4813]: I0317 09:42:00.550287 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562342-bq4l7" Mar 17 09:42:01 crc kubenswrapper[4813]: I0317 09:42:01.050563 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562342-bq4l7"] Mar 17 09:42:01 crc kubenswrapper[4813]: W0317 09:42:01.056099 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0b2f2eb2_09cd_422d_80a1_630d2811c390.slice/crio-23d03f3ceffc022c9b55280b8a345a2c08c4ab4ed2a8eeabc0d5571f1e735d66 WatchSource:0}: Error finding container 23d03f3ceffc022c9b55280b8a345a2c08c4ab4ed2a8eeabc0d5571f1e735d66: Status 404 returned error can't find the container with id 23d03f3ceffc022c9b55280b8a345a2c08c4ab4ed2a8eeabc0d5571f1e735d66 Mar 17 09:42:01 crc kubenswrapper[4813]: I0317 09:42:01.603274 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562342-bq4l7" event={"ID":"0b2f2eb2-09cd-422d-80a1-630d2811c390","Type":"ContainerStarted","Data":"23d03f3ceffc022c9b55280b8a345a2c08c4ab4ed2a8eeabc0d5571f1e735d66"} Mar 17 09:42:02 crc kubenswrapper[4813]: I0317 09:42:02.612839 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562342-bq4l7" event={"ID":"0b2f2eb2-09cd-422d-80a1-630d2811c390","Type":"ContainerStarted","Data":"1513f492fdb771fc3f67a5a814465f90982e5571d9d36fcd2ca7bdc8fed3a084"} Mar 17 09:42:02 crc kubenswrapper[4813]: I0317 09:42:02.637727 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562342-bq4l7" podStartSLOduration=1.59853171 podStartE2EDuration="2.637700254s" podCreationTimestamp="2026-03-17 09:42:00 +0000 UTC" firstStartedPulling="2026-03-17 09:42:01.058673167 +0000 UTC m=+1943.159476666" lastFinishedPulling="2026-03-17 09:42:02.097841671 +0000 UTC m=+1944.198645210" observedRunningTime="2026-03-17 09:42:02.627895369 +0000 UTC m=+1944.728698898" watchObservedRunningTime="2026-03-17 09:42:02.637700254 +0000 UTC m=+1944.738503763" Mar 17 09:42:03 crc kubenswrapper[4813]: I0317 09:42:03.637947 4813 generic.go:334] "Generic (PLEG): container finished" podID="0b2f2eb2-09cd-422d-80a1-630d2811c390" containerID="1513f492fdb771fc3f67a5a814465f90982e5571d9d36fcd2ca7bdc8fed3a084" exitCode=0 Mar 17 09:42:03 crc kubenswrapper[4813]: I0317 09:42:03.638033 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562342-bq4l7" event={"ID":"0b2f2eb2-09cd-422d-80a1-630d2811c390","Type":"ContainerDied","Data":"1513f492fdb771fc3f67a5a814465f90982e5571d9d36fcd2ca7bdc8fed3a084"} Mar 17 09:42:05 crc kubenswrapper[4813]: I0317 09:42:05.013195 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562342-bq4l7" Mar 17 09:42:05 crc kubenswrapper[4813]: I0317 09:42:05.176751 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgtcw\" (UniqueName: \"kubernetes.io/projected/0b2f2eb2-09cd-422d-80a1-630d2811c390-kube-api-access-bgtcw\") pod \"0b2f2eb2-09cd-422d-80a1-630d2811c390\" (UID: \"0b2f2eb2-09cd-422d-80a1-630d2811c390\") " Mar 17 09:42:05 crc kubenswrapper[4813]: I0317 09:42:05.182503 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b2f2eb2-09cd-422d-80a1-630d2811c390-kube-api-access-bgtcw" (OuterVolumeSpecName: "kube-api-access-bgtcw") pod "0b2f2eb2-09cd-422d-80a1-630d2811c390" (UID: "0b2f2eb2-09cd-422d-80a1-630d2811c390"). InnerVolumeSpecName "kube-api-access-bgtcw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:42:05 crc kubenswrapper[4813]: I0317 09:42:05.278844 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgtcw\" (UniqueName: \"kubernetes.io/projected/0b2f2eb2-09cd-422d-80a1-630d2811c390-kube-api-access-bgtcw\") on node \"crc\" DevicePath \"\"" Mar 17 09:42:05 crc kubenswrapper[4813]: I0317 09:42:05.663432 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562342-bq4l7" event={"ID":"0b2f2eb2-09cd-422d-80a1-630d2811c390","Type":"ContainerDied","Data":"23d03f3ceffc022c9b55280b8a345a2c08c4ab4ed2a8eeabc0d5571f1e735d66"} Mar 17 09:42:05 crc kubenswrapper[4813]: I0317 09:42:05.663492 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23d03f3ceffc022c9b55280b8a345a2c08c4ab4ed2a8eeabc0d5571f1e735d66" Mar 17 09:42:05 crc kubenswrapper[4813]: I0317 09:42:05.663537 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562342-bq4l7" Mar 17 09:42:05 crc kubenswrapper[4813]: I0317 09:42:05.717162 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562336-c6fds"] Mar 17 09:42:05 crc kubenswrapper[4813]: I0317 09:42:05.724580 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562336-c6fds"] Mar 17 09:42:06 crc kubenswrapper[4813]: I0317 09:42:06.751451 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0afc91f2-4193-40e2-9654-65a1b31326c3" path="/var/lib/kubelet/pods/0afc91f2-4193-40e2-9654-65a1b31326c3/volumes" Mar 17 09:42:15 crc kubenswrapper[4813]: I0317 09:42:15.790876 4813 generic.go:334] "Generic (PLEG): container finished" podID="63ad7386-248c-4dc8-a217-fb2cf3d4ef82" containerID="74f20b605ba23ca7e4078d252d3d9162a851d1b865f0427cee85376a5beac585" exitCode=0 Mar 17 09:42:15 crc kubenswrapper[4813]: I0317 09:42:15.790976 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-64lbt" event={"ID":"63ad7386-248c-4dc8-a217-fb2cf3d4ef82","Type":"ContainerDied","Data":"74f20b605ba23ca7e4078d252d3d9162a851d1b865f0427cee85376a5beac585"} Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.327585 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-64lbt" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.453443 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rnpp\" (UniqueName: \"kubernetes.io/projected/63ad7386-248c-4dc8-a217-fb2cf3d4ef82-kube-api-access-5rnpp\") pod \"63ad7386-248c-4dc8-a217-fb2cf3d4ef82\" (UID: \"63ad7386-248c-4dc8-a217-fb2cf3d4ef82\") " Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.454033 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/63ad7386-248c-4dc8-a217-fb2cf3d4ef82-ssh-key-openstack-edpm-ipam\") pod \"63ad7386-248c-4dc8-a217-fb2cf3d4ef82\" (UID: \"63ad7386-248c-4dc8-a217-fb2cf3d4ef82\") " Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.454222 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/63ad7386-248c-4dc8-a217-fb2cf3d4ef82-inventory\") pod \"63ad7386-248c-4dc8-a217-fb2cf3d4ef82\" (UID: \"63ad7386-248c-4dc8-a217-fb2cf3d4ef82\") " Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.460864 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/63ad7386-248c-4dc8-a217-fb2cf3d4ef82-kube-api-access-5rnpp" (OuterVolumeSpecName: "kube-api-access-5rnpp") pod "63ad7386-248c-4dc8-a217-fb2cf3d4ef82" (UID: "63ad7386-248c-4dc8-a217-fb2cf3d4ef82"). InnerVolumeSpecName "kube-api-access-5rnpp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.483310 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63ad7386-248c-4dc8-a217-fb2cf3d4ef82-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "63ad7386-248c-4dc8-a217-fb2cf3d4ef82" (UID: "63ad7386-248c-4dc8-a217-fb2cf3d4ef82"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.499436 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/63ad7386-248c-4dc8-a217-fb2cf3d4ef82-inventory" (OuterVolumeSpecName: "inventory") pod "63ad7386-248c-4dc8-a217-fb2cf3d4ef82" (UID: "63ad7386-248c-4dc8-a217-fb2cf3d4ef82"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.557559 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/63ad7386-248c-4dc8-a217-fb2cf3d4ef82-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.557634 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rnpp\" (UniqueName: \"kubernetes.io/projected/63ad7386-248c-4dc8-a217-fb2cf3d4ef82-kube-api-access-5rnpp\") on node \"crc\" DevicePath \"\"" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.557654 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/63ad7386-248c-4dc8-a217-fb2cf3d4ef82-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.818872 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-64lbt" event={"ID":"63ad7386-248c-4dc8-a217-fb2cf3d4ef82","Type":"ContainerDied","Data":"7d0011186339754aaedff52451f938a1cb9db2ddf81db1dea5f5756216a045e4"} Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.818969 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d0011186339754aaedff52451f938a1cb9db2ddf81db1dea5f5756216a045e4" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.818978 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-64lbt" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.941458 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5"] Mar 17 09:42:17 crc kubenswrapper[4813]: E0317 09:42:17.941970 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="63ad7386-248c-4dc8-a217-fb2cf3d4ef82" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.941996 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="63ad7386-248c-4dc8-a217-fb2cf3d4ef82" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 17 09:42:17 crc kubenswrapper[4813]: E0317 09:42:17.942014 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b2f2eb2-09cd-422d-80a1-630d2811c390" containerName="oc" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.942023 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b2f2eb2-09cd-422d-80a1-630d2811c390" containerName="oc" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.942266 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b2f2eb2-09cd-422d-80a1-630d2811c390" containerName="oc" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.942296 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="63ad7386-248c-4dc8-a217-fb2cf3d4ef82" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.943089 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.946466 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.947193 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.947727 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.952218 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hj9h4" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.966785 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-krpcf\" (UniqueName: \"kubernetes.io/projected/0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc-kube-api-access-krpcf\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5\" (UID: \"0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.967023 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5\" (UID: \"0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.967465 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5\" (UID: \"0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5" Mar 17 09:42:17 crc kubenswrapper[4813]: I0317 09:42:17.971502 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5"] Mar 17 09:42:18 crc kubenswrapper[4813]: I0317 09:42:18.069067 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5\" (UID: \"0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5" Mar 17 09:42:18 crc kubenswrapper[4813]: I0317 09:42:18.069257 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5\" (UID: \"0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5" Mar 17 09:42:18 crc kubenswrapper[4813]: I0317 09:42:18.069380 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-krpcf\" (UniqueName: \"kubernetes.io/projected/0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc-kube-api-access-krpcf\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5\" (UID: \"0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5" Mar 17 09:42:18 crc kubenswrapper[4813]: I0317 09:42:18.076791 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc-ssh-key-openstack-edpm-ipam\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5\" (UID: \"0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5" Mar 17 09:42:18 crc kubenswrapper[4813]: I0317 09:42:18.077117 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5\" (UID: \"0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5" Mar 17 09:42:18 crc kubenswrapper[4813]: I0317 09:42:18.101473 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-krpcf\" (UniqueName: \"kubernetes.io/projected/0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc-kube-api-access-krpcf\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5\" (UID: \"0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5" Mar 17 09:42:18 crc kubenswrapper[4813]: I0317 09:42:18.272925 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5" Mar 17 09:42:18 crc kubenswrapper[4813]: W0317 09:42:18.939158 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d8211b8_b1c4_4ec3_8dcd_5b7beefc7fdc.slice/crio-7cfe4e1aa714452356969e87fc51076682a731a4f48981f9a2f63719f9047324 WatchSource:0}: Error finding container 7cfe4e1aa714452356969e87fc51076682a731a4f48981f9a2f63719f9047324: Status 404 returned error can't find the container with id 7cfe4e1aa714452356969e87fc51076682a731a4f48981f9a2f63719f9047324 Mar 17 09:42:18 crc kubenswrapper[4813]: I0317 09:42:18.941957 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5"] Mar 17 09:42:19 crc kubenswrapper[4813]: I0317 09:42:19.848983 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5" event={"ID":"0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc","Type":"ContainerStarted","Data":"7cfe4e1aa714452356969e87fc51076682a731a4f48981f9a2f63719f9047324"} Mar 17 09:42:20 crc kubenswrapper[4813]: I0317 09:42:20.876525 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5" event={"ID":"0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc","Type":"ContainerStarted","Data":"1a599e642a711e83026623950696e369fd23c21b21949e1bd94d79f5922ec2a6"} Mar 17 09:42:20 crc kubenswrapper[4813]: I0317 09:42:20.911529 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5" podStartSLOduration=3.107067976 podStartE2EDuration="3.911508681s" podCreationTimestamp="2026-03-17 09:42:17 +0000 UTC" firstStartedPulling="2026-03-17 09:42:18.945806016 +0000 UTC m=+1961.046609515" lastFinishedPulling="2026-03-17 09:42:19.750246681 +0000 UTC m=+1961.851050220" observedRunningTime="2026-03-17 09:42:20.906824234 +0000 UTC m=+1963.007627753" watchObservedRunningTime="2026-03-17 09:42:20.911508681 +0000 UTC m=+1963.012312190" Mar 17 09:42:25 crc kubenswrapper[4813]: I0317 09:42:25.057333 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-pvfpm"] Mar 17 09:42:25 crc kubenswrapper[4813]: I0317 09:42:25.070455 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-f71a-account-create-update-mbsjs"] Mar 17 09:42:25 crc kubenswrapper[4813]: I0317 09:42:25.082649 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-f51b-account-create-update-sqrd9"] Mar 17 09:42:25 crc kubenswrapper[4813]: I0317 09:42:25.093099 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-vwt8q"] Mar 17 09:42:25 crc kubenswrapper[4813]: I0317 09:42:25.103368 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-pvfpm"] Mar 17 09:42:25 crc kubenswrapper[4813]: I0317 09:42:25.113425 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-8e34-account-create-update-s5l7t"] Mar 17 09:42:25 crc kubenswrapper[4813]: I0317 09:42:25.122446 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-f71a-account-create-update-mbsjs"] Mar 17 09:42:25 crc kubenswrapper[4813]: I0317 09:42:25.129983 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-n2hnd"] Mar 17 09:42:25 crc kubenswrapper[4813]: I0317 09:42:25.136948 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-vwt8q"] Mar 17 09:42:25 crc kubenswrapper[4813]: I0317 09:42:25.143456 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-f51b-account-create-update-sqrd9"] Mar 17 09:42:25 crc kubenswrapper[4813]: I0317 09:42:25.149729 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-8e34-account-create-update-s5l7t"] Mar 17 09:42:25 crc kubenswrapper[4813]: I0317 09:42:25.155794 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-n2hnd"] Mar 17 09:42:26 crc kubenswrapper[4813]: I0317 09:42:26.751189 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56ae9e04-8d01-4211-a2d0-2fc8dfff9593" path="/var/lib/kubelet/pods/56ae9e04-8d01-4211-a2d0-2fc8dfff9593/volumes" Mar 17 09:42:26 crc kubenswrapper[4813]: I0317 09:42:26.752844 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dfd5c4e-760a-47a9-a374-f457c63bc6fa" path="/var/lib/kubelet/pods/7dfd5c4e-760a-47a9-a374-f457c63bc6fa/volumes" Mar 17 09:42:26 crc kubenswrapper[4813]: I0317 09:42:26.754005 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b625cc1-7d38-4a27-b72c-f4ea768de618" path="/var/lib/kubelet/pods/8b625cc1-7d38-4a27-b72c-f4ea768de618/volumes" Mar 17 09:42:26 crc kubenswrapper[4813]: I0317 09:42:26.755255 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9534b041-bb7d-48c2-a5e1-ac9ca0af187c" path="/var/lib/kubelet/pods/9534b041-bb7d-48c2-a5e1-ac9ca0af187c/volumes" Mar 17 09:42:26 crc kubenswrapper[4813]: I0317 09:42:26.757637 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a06cbdf7-9def-426d-8057-35df25d255f6" path="/var/lib/kubelet/pods/a06cbdf7-9def-426d-8057-35df25d255f6/volumes" Mar 17 09:42:26 crc kubenswrapper[4813]: I0317 09:42:26.758830 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d67687bd-2db1-4a9d-8869-0aa9c453f5db" path="/var/lib/kubelet/pods/d67687bd-2db1-4a9d-8869-0aa9c453f5db/volumes" Mar 17 09:42:47 crc kubenswrapper[4813]: I0317 09:42:47.837166 4813 scope.go:117] "RemoveContainer" containerID="6be42a9854bd3c683909ea09eef0afa2b97a03c03b43cf65dade9d0f6b01af98" Mar 17 09:42:47 crc kubenswrapper[4813]: I0317 09:42:47.877207 4813 scope.go:117] "RemoveContainer" containerID="3103019bec4211362f28316017579ec261a49e21c63ac9db5f43fe82b5be0438" Mar 17 09:42:47 crc kubenswrapper[4813]: I0317 09:42:47.956660 4813 scope.go:117] "RemoveContainer" containerID="db0de2d2ef92ea95a2c95541079ed04a0eca3b0263df99907a6b1b18d39d3381" Mar 17 09:42:47 crc kubenswrapper[4813]: I0317 09:42:47.987335 4813 scope.go:117] "RemoveContainer" containerID="acf4ed9b63abf3ea698cbc23be53c44689ccf7056e109aef3577d28d27e86efb" Mar 17 09:42:48 crc kubenswrapper[4813]: I0317 09:42:48.020233 4813 scope.go:117] "RemoveContainer" containerID="d58cc6f0d9a5c4bb85d3259485359f7fe7e20d5f74705fe50c90a3f040deae9c" Mar 17 09:42:48 crc kubenswrapper[4813]: I0317 09:42:48.088811 4813 scope.go:117] "RemoveContainer" containerID="10d6ff5ffda0d5731232121b139d9ec7456c86e1ff460c20c0860d546c309749" Mar 17 09:42:48 crc kubenswrapper[4813]: I0317 09:42:48.118172 4813 scope.go:117] "RemoveContainer" containerID="e9fd1c7b79e5b326c1264811d9cb8729ac74fed9395c7bced7ae92947a232cdb" Mar 17 09:42:51 crc kubenswrapper[4813]: I0317 09:42:51.062469 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jdtxs"] Mar 17 09:42:51 crc kubenswrapper[4813]: I0317 09:42:51.077551 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-jdtxs"] Mar 17 09:42:52 crc kubenswrapper[4813]: I0317 09:42:52.745677 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48793e32-107b-4ba7-b3a6-3e8172306408" path="/var/lib/kubelet/pods/48793e32-107b-4ba7-b3a6-3e8172306408/volumes" Mar 17 09:43:12 crc kubenswrapper[4813]: I0317 09:43:12.052517 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vqcqs"] Mar 17 09:43:12 crc kubenswrapper[4813]: I0317 09:43:12.067383 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-vqcqs"] Mar 17 09:43:12 crc kubenswrapper[4813]: I0317 09:43:12.743518 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="500133a3-3767-4bd8-8011-066d1b77ddca" path="/var/lib/kubelet/pods/500133a3-3767-4bd8-8011-066d1b77ddca/volumes" Mar 17 09:43:13 crc kubenswrapper[4813]: I0317 09:43:13.052525 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-2j8vj"] Mar 17 09:43:13 crc kubenswrapper[4813]: I0317 09:43:13.067049 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-2j8vj"] Mar 17 09:43:13 crc kubenswrapper[4813]: I0317 09:43:13.482835 4813 generic.go:334] "Generic (PLEG): container finished" podID="0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc" containerID="1a599e642a711e83026623950696e369fd23c21b21949e1bd94d79f5922ec2a6" exitCode=0 Mar 17 09:43:13 crc kubenswrapper[4813]: I0317 09:43:13.482896 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5" event={"ID":"0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc","Type":"ContainerDied","Data":"1a599e642a711e83026623950696e369fd23c21b21949e1bd94d79f5922ec2a6"} Mar 17 09:43:14 crc kubenswrapper[4813]: I0317 09:43:14.755846 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d08fdcd8-7f6c-468e-bc05-5f0c8a4add79" path="/var/lib/kubelet/pods/d08fdcd8-7f6c-468e-bc05-5f0c8a4add79/volumes" Mar 17 09:43:14 crc kubenswrapper[4813]: I0317 09:43:14.923190 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5" Mar 17 09:43:14 crc kubenswrapper[4813]: I0317 09:43:14.987461 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-krpcf\" (UniqueName: \"kubernetes.io/projected/0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc-kube-api-access-krpcf\") pod \"0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc\" (UID: \"0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc\") " Mar 17 09:43:14 crc kubenswrapper[4813]: I0317 09:43:14.987620 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc-inventory\") pod \"0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc\" (UID: \"0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc\") " Mar 17 09:43:14 crc kubenswrapper[4813]: I0317 09:43:14.987666 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc-ssh-key-openstack-edpm-ipam\") pod \"0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc\" (UID: \"0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc\") " Mar 17 09:43:14 crc kubenswrapper[4813]: I0317 09:43:14.993309 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc-kube-api-access-krpcf" (OuterVolumeSpecName: "kube-api-access-krpcf") pod "0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc" (UID: "0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc"). InnerVolumeSpecName "kube-api-access-krpcf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.015880 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc-inventory" (OuterVolumeSpecName: "inventory") pod "0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc" (UID: "0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.017639 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc" (UID: "0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.090002 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-krpcf\" (UniqueName: \"kubernetes.io/projected/0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc-kube-api-access-krpcf\") on node \"crc\" DevicePath \"\"" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.090250 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.090261 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.504734 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5" event={"ID":"0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc","Type":"ContainerDied","Data":"7cfe4e1aa714452356969e87fc51076682a731a4f48981f9a2f63719f9047324"} Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.504789 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.504799 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7cfe4e1aa714452356969e87fc51076682a731a4f48981f9a2f63719f9047324" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.648209 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ht9q9"] Mar 17 09:43:15 crc kubenswrapper[4813]: E0317 09:43:15.648715 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.648733 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.648964 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.649620 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ht9q9" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.652091 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hj9h4" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.652967 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.653763 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.654043 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.659690 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ht9q9"] Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.706947 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wxb6c\" (UniqueName: \"kubernetes.io/projected/1d6a2187-6711-4b3f-988a-845c5b7e18f7-kube-api-access-wxb6c\") pod \"ssh-known-hosts-edpm-deployment-ht9q9\" (UID: \"1d6a2187-6711-4b3f-988a-845c5b7e18f7\") " pod="openstack/ssh-known-hosts-edpm-deployment-ht9q9" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.707059 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1d6a2187-6711-4b3f-988a-845c5b7e18f7-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ht9q9\" (UID: \"1d6a2187-6711-4b3f-988a-845c5b7e18f7\") " pod="openstack/ssh-known-hosts-edpm-deployment-ht9q9" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.707094 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1d6a2187-6711-4b3f-988a-845c5b7e18f7-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ht9q9\" (UID: \"1d6a2187-6711-4b3f-988a-845c5b7e18f7\") " pod="openstack/ssh-known-hosts-edpm-deployment-ht9q9" Mar 17 09:43:15 crc kubenswrapper[4813]: E0317 09:43:15.733642 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0d8211b8_b1c4_4ec3_8dcd_5b7beefc7fdc.slice\": RecentStats: unable to find data in memory cache]" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.810039 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1d6a2187-6711-4b3f-988a-845c5b7e18f7-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ht9q9\" (UID: \"1d6a2187-6711-4b3f-988a-845c5b7e18f7\") " pod="openstack/ssh-known-hosts-edpm-deployment-ht9q9" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.810136 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1d6a2187-6711-4b3f-988a-845c5b7e18f7-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ht9q9\" (UID: \"1d6a2187-6711-4b3f-988a-845c5b7e18f7\") " pod="openstack/ssh-known-hosts-edpm-deployment-ht9q9" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.810433 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wxb6c\" (UniqueName: \"kubernetes.io/projected/1d6a2187-6711-4b3f-988a-845c5b7e18f7-kube-api-access-wxb6c\") pod \"ssh-known-hosts-edpm-deployment-ht9q9\" (UID: \"1d6a2187-6711-4b3f-988a-845c5b7e18f7\") " pod="openstack/ssh-known-hosts-edpm-deployment-ht9q9" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.814646 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1d6a2187-6711-4b3f-988a-845c5b7e18f7-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-ht9q9\" (UID: \"1d6a2187-6711-4b3f-988a-845c5b7e18f7\") " pod="openstack/ssh-known-hosts-edpm-deployment-ht9q9" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.822833 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1d6a2187-6711-4b3f-988a-845c5b7e18f7-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-ht9q9\" (UID: \"1d6a2187-6711-4b3f-988a-845c5b7e18f7\") " pod="openstack/ssh-known-hosts-edpm-deployment-ht9q9" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.864635 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wxb6c\" (UniqueName: \"kubernetes.io/projected/1d6a2187-6711-4b3f-988a-845c5b7e18f7-kube-api-access-wxb6c\") pod \"ssh-known-hosts-edpm-deployment-ht9q9\" (UID: \"1d6a2187-6711-4b3f-988a-845c5b7e18f7\") " pod="openstack/ssh-known-hosts-edpm-deployment-ht9q9" Mar 17 09:43:15 crc kubenswrapper[4813]: I0317 09:43:15.975006 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ht9q9" Mar 17 09:43:16 crc kubenswrapper[4813]: I0317 09:43:16.544045 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-ht9q9"] Mar 17 09:43:17 crc kubenswrapper[4813]: I0317 09:43:17.525817 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ht9q9" event={"ID":"1d6a2187-6711-4b3f-988a-845c5b7e18f7","Type":"ContainerStarted","Data":"396fca29dd714b3efba986e3cefdf57a22e3ec0e110ced5423fbef3f4a0d001c"} Mar 17 09:43:17 crc kubenswrapper[4813]: I0317 09:43:17.526177 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ht9q9" event={"ID":"1d6a2187-6711-4b3f-988a-845c5b7e18f7","Type":"ContainerStarted","Data":"c11428ac54024db698be5b8389e67afa9c6d40ea66c2b18a9bccd18107ecbbd3"} Mar 17 09:43:17 crc kubenswrapper[4813]: I0317 09:43:17.564107 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-ht9q9" podStartSLOduration=2.16325629 podStartE2EDuration="2.564072119s" podCreationTimestamp="2026-03-17 09:43:15 +0000 UTC" firstStartedPulling="2026-03-17 09:43:16.544773445 +0000 UTC m=+2018.645576954" lastFinishedPulling="2026-03-17 09:43:16.945589284 +0000 UTC m=+2019.046392783" observedRunningTime="2026-03-17 09:43:17.550821015 +0000 UTC m=+2019.651624554" watchObservedRunningTime="2026-03-17 09:43:17.564072119 +0000 UTC m=+2019.664875658" Mar 17 09:43:24 crc kubenswrapper[4813]: I0317 09:43:24.595186 4813 generic.go:334] "Generic (PLEG): container finished" podID="1d6a2187-6711-4b3f-988a-845c5b7e18f7" containerID="396fca29dd714b3efba986e3cefdf57a22e3ec0e110ced5423fbef3f4a0d001c" exitCode=0 Mar 17 09:43:24 crc kubenswrapper[4813]: I0317 09:43:24.595290 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ht9q9" event={"ID":"1d6a2187-6711-4b3f-988a-845c5b7e18f7","Type":"ContainerDied","Data":"396fca29dd714b3efba986e3cefdf57a22e3ec0e110ced5423fbef3f4a0d001c"} Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.190685 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ht9q9" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.347393 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1d6a2187-6711-4b3f-988a-845c5b7e18f7-ssh-key-openstack-edpm-ipam\") pod \"1d6a2187-6711-4b3f-988a-845c5b7e18f7\" (UID: \"1d6a2187-6711-4b3f-988a-845c5b7e18f7\") " Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.347557 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxb6c\" (UniqueName: \"kubernetes.io/projected/1d6a2187-6711-4b3f-988a-845c5b7e18f7-kube-api-access-wxb6c\") pod \"1d6a2187-6711-4b3f-988a-845c5b7e18f7\" (UID: \"1d6a2187-6711-4b3f-988a-845c5b7e18f7\") " Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.347772 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1d6a2187-6711-4b3f-988a-845c5b7e18f7-inventory-0\") pod \"1d6a2187-6711-4b3f-988a-845c5b7e18f7\" (UID: \"1d6a2187-6711-4b3f-988a-845c5b7e18f7\") " Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.353975 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d6a2187-6711-4b3f-988a-845c5b7e18f7-kube-api-access-wxb6c" (OuterVolumeSpecName: "kube-api-access-wxb6c") pod "1d6a2187-6711-4b3f-988a-845c5b7e18f7" (UID: "1d6a2187-6711-4b3f-988a-845c5b7e18f7"). InnerVolumeSpecName "kube-api-access-wxb6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.380674 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d6a2187-6711-4b3f-988a-845c5b7e18f7-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "1d6a2187-6711-4b3f-988a-845c5b7e18f7" (UID: "1d6a2187-6711-4b3f-988a-845c5b7e18f7"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.399585 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d6a2187-6711-4b3f-988a-845c5b7e18f7-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "1d6a2187-6711-4b3f-988a-845c5b7e18f7" (UID: "1d6a2187-6711-4b3f-988a-845c5b7e18f7"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.450212 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxb6c\" (UniqueName: \"kubernetes.io/projected/1d6a2187-6711-4b3f-988a-845c5b7e18f7-kube-api-access-wxb6c\") on node \"crc\" DevicePath \"\"" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.450247 4813 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/1d6a2187-6711-4b3f-988a-845c5b7e18f7-inventory-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.450258 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/1d6a2187-6711-4b3f-988a-845c5b7e18f7-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.617388 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-ht9q9" event={"ID":"1d6a2187-6711-4b3f-988a-845c5b7e18f7","Type":"ContainerDied","Data":"c11428ac54024db698be5b8389e67afa9c6d40ea66c2b18a9bccd18107ecbbd3"} Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.617454 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c11428ac54024db698be5b8389e67afa9c6d40ea66c2b18a9bccd18107ecbbd3" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.617484 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-ht9q9" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.710574 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jhl6m"] Mar 17 09:43:26 crc kubenswrapper[4813]: E0317 09:43:26.711405 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d6a2187-6711-4b3f-988a-845c5b7e18f7" containerName="ssh-known-hosts-edpm-deployment" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.711427 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d6a2187-6711-4b3f-988a-845c5b7e18f7" containerName="ssh-known-hosts-edpm-deployment" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.711695 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d6a2187-6711-4b3f-988a-845c5b7e18f7" containerName="ssh-known-hosts-edpm-deployment" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.712405 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jhl6m" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.714844 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.715538 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.715746 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.716341 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hj9h4" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.746860 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jhl6m"] Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.862367 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/540d96f7-a4a1-490f-9c57-997a2c6dc9ab-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jhl6m\" (UID: \"540d96f7-a4a1-490f-9c57-997a2c6dc9ab\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jhl6m" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.862452 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/540d96f7-a4a1-490f-9c57-997a2c6dc9ab-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jhl6m\" (UID: \"540d96f7-a4a1-490f-9c57-997a2c6dc9ab\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jhl6m" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.862610 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spx97\" (UniqueName: \"kubernetes.io/projected/540d96f7-a4a1-490f-9c57-997a2c6dc9ab-kube-api-access-spx97\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jhl6m\" (UID: \"540d96f7-a4a1-490f-9c57-997a2c6dc9ab\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jhl6m" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.964922 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spx97\" (UniqueName: \"kubernetes.io/projected/540d96f7-a4a1-490f-9c57-997a2c6dc9ab-kube-api-access-spx97\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jhl6m\" (UID: \"540d96f7-a4a1-490f-9c57-997a2c6dc9ab\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jhl6m" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.965100 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/540d96f7-a4a1-490f-9c57-997a2c6dc9ab-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jhl6m\" (UID: \"540d96f7-a4a1-490f-9c57-997a2c6dc9ab\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jhl6m" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.965231 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/540d96f7-a4a1-490f-9c57-997a2c6dc9ab-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jhl6m\" (UID: \"540d96f7-a4a1-490f-9c57-997a2c6dc9ab\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jhl6m" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.973848 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/540d96f7-a4a1-490f-9c57-997a2c6dc9ab-ssh-key-openstack-edpm-ipam\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jhl6m\" (UID: \"540d96f7-a4a1-490f-9c57-997a2c6dc9ab\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jhl6m" Mar 17 09:43:26 crc kubenswrapper[4813]: I0317 09:43:26.974265 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/540d96f7-a4a1-490f-9c57-997a2c6dc9ab-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jhl6m\" (UID: \"540d96f7-a4a1-490f-9c57-997a2c6dc9ab\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jhl6m" Mar 17 09:43:27 crc kubenswrapper[4813]: I0317 09:43:27.001347 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spx97\" (UniqueName: \"kubernetes.io/projected/540d96f7-a4a1-490f-9c57-997a2c6dc9ab-kube-api-access-spx97\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jhl6m\" (UID: \"540d96f7-a4a1-490f-9c57-997a2c6dc9ab\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jhl6m" Mar 17 09:43:27 crc kubenswrapper[4813]: I0317 09:43:27.035035 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jhl6m" Mar 17 09:43:27 crc kubenswrapper[4813]: I0317 09:43:27.435524 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jhl6m"] Mar 17 09:43:27 crc kubenswrapper[4813]: W0317 09:43:27.451510 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod540d96f7_a4a1_490f_9c57_997a2c6dc9ab.slice/crio-2d411e86aa70eaad0d5a983221929bf644547d65a88472ec7f67138ca083f288 WatchSource:0}: Error finding container 2d411e86aa70eaad0d5a983221929bf644547d65a88472ec7f67138ca083f288: Status 404 returned error can't find the container with id 2d411e86aa70eaad0d5a983221929bf644547d65a88472ec7f67138ca083f288 Mar 17 09:43:27 crc kubenswrapper[4813]: I0317 09:43:27.629173 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jhl6m" event={"ID":"540d96f7-a4a1-490f-9c57-997a2c6dc9ab","Type":"ContainerStarted","Data":"2d411e86aa70eaad0d5a983221929bf644547d65a88472ec7f67138ca083f288"} Mar 17 09:43:28 crc kubenswrapper[4813]: I0317 09:43:28.644907 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jhl6m" event={"ID":"540d96f7-a4a1-490f-9c57-997a2c6dc9ab","Type":"ContainerStarted","Data":"2228a8839ed4e943e0a2060234d69063ba83cc436b5ef38cee48505af3530f87"} Mar 17 09:43:28 crc kubenswrapper[4813]: I0317 09:43:28.688820 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jhl6m" podStartSLOduration=1.968523712 podStartE2EDuration="2.688800852s" podCreationTimestamp="2026-03-17 09:43:26 +0000 UTC" firstStartedPulling="2026-03-17 09:43:27.455718111 +0000 UTC m=+2029.556521620" lastFinishedPulling="2026-03-17 09:43:28.175995241 +0000 UTC m=+2030.276798760" observedRunningTime="2026-03-17 09:43:28.672154113 +0000 UTC m=+2030.772957612" watchObservedRunningTime="2026-03-17 09:43:28.688800852 +0000 UTC m=+2030.789604361" Mar 17 09:43:36 crc kubenswrapper[4813]: I0317 09:43:36.736692 4813 generic.go:334] "Generic (PLEG): container finished" podID="540d96f7-a4a1-490f-9c57-997a2c6dc9ab" containerID="2228a8839ed4e943e0a2060234d69063ba83cc436b5ef38cee48505af3530f87" exitCode=0 Mar 17 09:43:36 crc kubenswrapper[4813]: I0317 09:43:36.747004 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jhl6m" event={"ID":"540d96f7-a4a1-490f-9c57-997a2c6dc9ab","Type":"ContainerDied","Data":"2228a8839ed4e943e0a2060234d69063ba83cc436b5ef38cee48505af3530f87"} Mar 17 09:43:38 crc kubenswrapper[4813]: I0317 09:43:38.223061 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jhl6m" Mar 17 09:43:38 crc kubenswrapper[4813]: I0317 09:43:38.320028 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/540d96f7-a4a1-490f-9c57-997a2c6dc9ab-ssh-key-openstack-edpm-ipam\") pod \"540d96f7-a4a1-490f-9c57-997a2c6dc9ab\" (UID: \"540d96f7-a4a1-490f-9c57-997a2c6dc9ab\") " Mar 17 09:43:38 crc kubenswrapper[4813]: I0317 09:43:38.320091 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spx97\" (UniqueName: \"kubernetes.io/projected/540d96f7-a4a1-490f-9c57-997a2c6dc9ab-kube-api-access-spx97\") pod \"540d96f7-a4a1-490f-9c57-997a2c6dc9ab\" (UID: \"540d96f7-a4a1-490f-9c57-997a2c6dc9ab\") " Mar 17 09:43:38 crc kubenswrapper[4813]: I0317 09:43:38.320314 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/540d96f7-a4a1-490f-9c57-997a2c6dc9ab-inventory\") pod \"540d96f7-a4a1-490f-9c57-997a2c6dc9ab\" (UID: \"540d96f7-a4a1-490f-9c57-997a2c6dc9ab\") " Mar 17 09:43:38 crc kubenswrapper[4813]: I0317 09:43:38.327192 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/540d96f7-a4a1-490f-9c57-997a2c6dc9ab-kube-api-access-spx97" (OuterVolumeSpecName: "kube-api-access-spx97") pod "540d96f7-a4a1-490f-9c57-997a2c6dc9ab" (UID: "540d96f7-a4a1-490f-9c57-997a2c6dc9ab"). InnerVolumeSpecName "kube-api-access-spx97". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:43:38 crc kubenswrapper[4813]: I0317 09:43:38.366539 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/540d96f7-a4a1-490f-9c57-997a2c6dc9ab-inventory" (OuterVolumeSpecName: "inventory") pod "540d96f7-a4a1-490f-9c57-997a2c6dc9ab" (UID: "540d96f7-a4a1-490f-9c57-997a2c6dc9ab"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:43:38 crc kubenswrapper[4813]: I0317 09:43:38.371277 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/540d96f7-a4a1-490f-9c57-997a2c6dc9ab-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "540d96f7-a4a1-490f-9c57-997a2c6dc9ab" (UID: "540d96f7-a4a1-490f-9c57-997a2c6dc9ab"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:43:38 crc kubenswrapper[4813]: I0317 09:43:38.422660 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spx97\" (UniqueName: \"kubernetes.io/projected/540d96f7-a4a1-490f-9c57-997a2c6dc9ab-kube-api-access-spx97\") on node \"crc\" DevicePath \"\"" Mar 17 09:43:38 crc kubenswrapper[4813]: I0317 09:43:38.422841 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/540d96f7-a4a1-490f-9c57-997a2c6dc9ab-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 09:43:38 crc kubenswrapper[4813]: I0317 09:43:38.422958 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/540d96f7-a4a1-490f-9c57-997a2c6dc9ab-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 09:43:38 crc kubenswrapper[4813]: I0317 09:43:38.764560 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jhl6m" event={"ID":"540d96f7-a4a1-490f-9c57-997a2c6dc9ab","Type":"ContainerDied","Data":"2d411e86aa70eaad0d5a983221929bf644547d65a88472ec7f67138ca083f288"} Mar 17 09:43:38 crc kubenswrapper[4813]: I0317 09:43:38.764677 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d411e86aa70eaad0d5a983221929bf644547d65a88472ec7f67138ca083f288" Mar 17 09:43:38 crc kubenswrapper[4813]: I0317 09:43:38.764694 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jhl6m" Mar 17 09:43:39 crc kubenswrapper[4813]: I0317 09:43:39.035141 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk"] Mar 17 09:43:39 crc kubenswrapper[4813]: E0317 09:43:39.035647 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="540d96f7-a4a1-490f-9c57-997a2c6dc9ab" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 17 09:43:39 crc kubenswrapper[4813]: I0317 09:43:39.035669 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="540d96f7-a4a1-490f-9c57-997a2c6dc9ab" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 17 09:43:39 crc kubenswrapper[4813]: I0317 09:43:39.035943 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="540d96f7-a4a1-490f-9c57-997a2c6dc9ab" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Mar 17 09:43:39 crc kubenswrapper[4813]: I0317 09:43:39.036731 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk" Mar 17 09:43:39 crc kubenswrapper[4813]: I0317 09:43:39.040892 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 09:43:39 crc kubenswrapper[4813]: I0317 09:43:39.041184 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hj9h4" Mar 17 09:43:39 crc kubenswrapper[4813]: I0317 09:43:39.043191 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 09:43:39 crc kubenswrapper[4813]: I0317 09:43:39.046443 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk"] Mar 17 09:43:39 crc kubenswrapper[4813]: I0317 09:43:39.056166 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 09:43:39 crc kubenswrapper[4813]: I0317 09:43:39.138100 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/60339fa3-6329-4aa7-a958-d28be7f562cc-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk\" (UID: \"60339fa3-6329-4aa7-a958-d28be7f562cc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk" Mar 17 09:43:39 crc kubenswrapper[4813]: I0317 09:43:39.138154 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nngc5\" (UniqueName: \"kubernetes.io/projected/60339fa3-6329-4aa7-a958-d28be7f562cc-kube-api-access-nngc5\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk\" (UID: \"60339fa3-6329-4aa7-a958-d28be7f562cc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk" Mar 17 09:43:39 crc kubenswrapper[4813]: I0317 09:43:39.138478 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60339fa3-6329-4aa7-a958-d28be7f562cc-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk\" (UID: \"60339fa3-6329-4aa7-a958-d28be7f562cc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk" Mar 17 09:43:39 crc kubenswrapper[4813]: I0317 09:43:39.241110 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/60339fa3-6329-4aa7-a958-d28be7f562cc-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk\" (UID: \"60339fa3-6329-4aa7-a958-d28be7f562cc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk" Mar 17 09:43:39 crc kubenswrapper[4813]: I0317 09:43:39.241194 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nngc5\" (UniqueName: \"kubernetes.io/projected/60339fa3-6329-4aa7-a958-d28be7f562cc-kube-api-access-nngc5\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk\" (UID: \"60339fa3-6329-4aa7-a958-d28be7f562cc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk" Mar 17 09:43:39 crc kubenswrapper[4813]: I0317 09:43:39.241305 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60339fa3-6329-4aa7-a958-d28be7f562cc-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk\" (UID: \"60339fa3-6329-4aa7-a958-d28be7f562cc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk" Mar 17 09:43:39 crc kubenswrapper[4813]: I0317 09:43:39.248871 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/60339fa3-6329-4aa7-a958-d28be7f562cc-ssh-key-openstack-edpm-ipam\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk\" (UID: \"60339fa3-6329-4aa7-a958-d28be7f562cc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk" Mar 17 09:43:39 crc kubenswrapper[4813]: I0317 09:43:39.256803 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60339fa3-6329-4aa7-a958-d28be7f562cc-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk\" (UID: \"60339fa3-6329-4aa7-a958-d28be7f562cc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk" Mar 17 09:43:39 crc kubenswrapper[4813]: I0317 09:43:39.281239 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nngc5\" (UniqueName: \"kubernetes.io/projected/60339fa3-6329-4aa7-a958-d28be7f562cc-kube-api-access-nngc5\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk\" (UID: \"60339fa3-6329-4aa7-a958-d28be7f562cc\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk" Mar 17 09:43:39 crc kubenswrapper[4813]: I0317 09:43:39.365331 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk" Mar 17 09:43:39 crc kubenswrapper[4813]: I0317 09:43:39.977521 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk"] Mar 17 09:43:40 crc kubenswrapper[4813]: I0317 09:43:40.781511 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk" event={"ID":"60339fa3-6329-4aa7-a958-d28be7f562cc","Type":"ContainerStarted","Data":"fe7bab7ef69819ecf6dc0265295ab6782fc1e249f0c9dc69b4a4a5c63e6fcaba"} Mar 17 09:43:42 crc kubenswrapper[4813]: I0317 09:43:42.807341 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk" event={"ID":"60339fa3-6329-4aa7-a958-d28be7f562cc","Type":"ContainerStarted","Data":"03b324705eb8cc9285a7111e309ec2e5108a2a3b6a0619e601387baf4c3ce571"} Mar 17 09:43:42 crc kubenswrapper[4813]: I0317 09:43:42.844668 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk" podStartSLOduration=1.839208835 podStartE2EDuration="3.844640959s" podCreationTimestamp="2026-03-17 09:43:39 +0000 UTC" firstStartedPulling="2026-03-17 09:43:39.979589941 +0000 UTC m=+2042.080393440" lastFinishedPulling="2026-03-17 09:43:41.985022035 +0000 UTC m=+2044.085825564" observedRunningTime="2026-03-17 09:43:42.835857706 +0000 UTC m=+2044.936661245" watchObservedRunningTime="2026-03-17 09:43:42.844640959 +0000 UTC m=+2044.945444498" Mar 17 09:43:48 crc kubenswrapper[4813]: I0317 09:43:48.269975 4813 scope.go:117] "RemoveContainer" containerID="ebdf2c0cb40a2d460d25104c47216d741f61c3ee45bfae8c282b3ef6465331d1" Mar 17 09:43:48 crc kubenswrapper[4813]: I0317 09:43:48.349816 4813 scope.go:117] "RemoveContainer" containerID="3671bbe1c1f3b2c6bde24a80e90ff0a32ab448a617d0b0d5858aeb5ce6fe5c63" Mar 17 09:43:48 crc kubenswrapper[4813]: I0317 09:43:48.427716 4813 scope.go:117] "RemoveContainer" containerID="cd051cb78eab1e8bb6bdda3cac209d8aa12d7ecdf22cc87948d4018db9a49829" Mar 17 09:43:52 crc kubenswrapper[4813]: I0317 09:43:52.092977 4813 generic.go:334] "Generic (PLEG): container finished" podID="60339fa3-6329-4aa7-a958-d28be7f562cc" containerID="03b324705eb8cc9285a7111e309ec2e5108a2a3b6a0619e601387baf4c3ce571" exitCode=0 Mar 17 09:43:52 crc kubenswrapper[4813]: I0317 09:43:52.093306 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk" event={"ID":"60339fa3-6329-4aa7-a958-d28be7f562cc","Type":"ContainerDied","Data":"03b324705eb8cc9285a7111e309ec2e5108a2a3b6a0619e601387baf4c3ce571"} Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.043627 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.115591 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk" event={"ID":"60339fa3-6329-4aa7-a958-d28be7f562cc","Type":"ContainerDied","Data":"fe7bab7ef69819ecf6dc0265295ab6782fc1e249f0c9dc69b4a4a5c63e6fcaba"} Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.115638 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.115649 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe7bab7ef69819ecf6dc0265295ab6782fc1e249f0c9dc69b4a4a5c63e6fcaba" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.146695 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/60339fa3-6329-4aa7-a958-d28be7f562cc-ssh-key-openstack-edpm-ipam\") pod \"60339fa3-6329-4aa7-a958-d28be7f562cc\" (UID: \"60339fa3-6329-4aa7-a958-d28be7f562cc\") " Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.146802 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nngc5\" (UniqueName: \"kubernetes.io/projected/60339fa3-6329-4aa7-a958-d28be7f562cc-kube-api-access-nngc5\") pod \"60339fa3-6329-4aa7-a958-d28be7f562cc\" (UID: \"60339fa3-6329-4aa7-a958-d28be7f562cc\") " Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.146938 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60339fa3-6329-4aa7-a958-d28be7f562cc-inventory\") pod \"60339fa3-6329-4aa7-a958-d28be7f562cc\" (UID: \"60339fa3-6329-4aa7-a958-d28be7f562cc\") " Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.160706 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60339fa3-6329-4aa7-a958-d28be7f562cc-kube-api-access-nngc5" (OuterVolumeSpecName: "kube-api-access-nngc5") pod "60339fa3-6329-4aa7-a958-d28be7f562cc" (UID: "60339fa3-6329-4aa7-a958-d28be7f562cc"). InnerVolumeSpecName "kube-api-access-nngc5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.181739 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60339fa3-6329-4aa7-a958-d28be7f562cc-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "60339fa3-6329-4aa7-a958-d28be7f562cc" (UID: "60339fa3-6329-4aa7-a958-d28be7f562cc"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.211319 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60339fa3-6329-4aa7-a958-d28be7f562cc-inventory" (OuterVolumeSpecName: "inventory") pod "60339fa3-6329-4aa7-a958-d28be7f562cc" (UID: "60339fa3-6329-4aa7-a958-d28be7f562cc"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.224861 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt"] Mar 17 09:43:54 crc kubenswrapper[4813]: E0317 09:43:54.233522 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60339fa3-6329-4aa7-a958-d28be7f562cc" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.233569 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="60339fa3-6329-4aa7-a958-d28be7f562cc" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.234724 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="60339fa3-6329-4aa7-a958-d28be7f562cc" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.245121 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt"] Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.245250 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.248472 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.248572 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.248809 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.249006 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.251039 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/60339fa3-6329-4aa7-a958-d28be7f562cc-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.254045 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nngc5\" (UniqueName: \"kubernetes.io/projected/60339fa3-6329-4aa7-a958-d28be7f562cc-kube-api-access-nngc5\") on node \"crc\" DevicePath \"\"" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.254129 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/60339fa3-6329-4aa7-a958-d28be7f562cc-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.355985 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.356258 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.356282 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.356299 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.356339 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.356356 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.356378 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.356413 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.356436 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcz9d\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-kube-api-access-lcz9d\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.356457 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.356483 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.356507 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.356547 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.356617 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.458222 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.458493 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcz9d\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-kube-api-access-lcz9d\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.458629 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.459238 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.459404 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.459537 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.459731 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.459878 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.459997 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.460092 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.460188 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.460321 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.460424 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.460544 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.464625 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-ssh-key-openstack-edpm-ipam\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.464664 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.465399 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.465428 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.466083 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.466428 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.467202 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.467253 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.467525 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.467961 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.468489 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.470495 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.474595 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.475190 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcz9d\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-kube-api-access-lcz9d\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:54 crc kubenswrapper[4813]: I0317 09:43:54.584220 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:43:55 crc kubenswrapper[4813]: I0317 09:43:55.105171 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt"] Mar 17 09:43:55 crc kubenswrapper[4813]: I0317 09:43:55.126118 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" event={"ID":"d8cecca4-b584-4537-810f-02c1818d2e99","Type":"ContainerStarted","Data":"b0eef6fbe7476b2fe451ddd27c0dd9df52d3e0ac26264e67de329bb648d923f8"} Mar 17 09:43:56 crc kubenswrapper[4813]: I0317 09:43:56.138859 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" event={"ID":"d8cecca4-b584-4537-810f-02c1818d2e99","Type":"ContainerStarted","Data":"38b171d4c8baab1dd27791ffc6e04e142a9035567145de55c14c6198535f264f"} Mar 17 09:43:56 crc kubenswrapper[4813]: I0317 09:43:56.171791 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" podStartSLOduration=1.7447515930000002 podStartE2EDuration="2.171759988s" podCreationTimestamp="2026-03-17 09:43:54 +0000 UTC" firstStartedPulling="2026-03-17 09:43:55.103425045 +0000 UTC m=+2057.204228554" lastFinishedPulling="2026-03-17 09:43:55.53043341 +0000 UTC m=+2057.631236949" observedRunningTime="2026-03-17 09:43:56.171426717 +0000 UTC m=+2058.272230256" watchObservedRunningTime="2026-03-17 09:43:56.171759988 +0000 UTC m=+2058.272563487" Mar 17 09:43:58 crc kubenswrapper[4813]: I0317 09:43:58.065134 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-v8vll"] Mar 17 09:43:58 crc kubenswrapper[4813]: I0317 09:43:58.078232 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-v8vll"] Mar 17 09:43:58 crc kubenswrapper[4813]: I0317 09:43:58.751945 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6607c0ac-29d9-4cd9-9d95-9cfce3717c76" path="/var/lib/kubelet/pods/6607c0ac-29d9-4cd9-9d95-9cfce3717c76/volumes" Mar 17 09:44:00 crc kubenswrapper[4813]: I0317 09:44:00.144178 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562344-6kk6q"] Mar 17 09:44:00 crc kubenswrapper[4813]: I0317 09:44:00.145905 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562344-6kk6q" Mar 17 09:44:00 crc kubenswrapper[4813]: I0317 09:44:00.148700 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:44:00 crc kubenswrapper[4813]: I0317 09:44:00.149225 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:44:00 crc kubenswrapper[4813]: I0317 09:44:00.149996 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:44:00 crc kubenswrapper[4813]: I0317 09:44:00.159445 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562344-6kk6q"] Mar 17 09:44:00 crc kubenswrapper[4813]: I0317 09:44:00.205572 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnsq5\" (UniqueName: \"kubernetes.io/projected/a530431d-cc3a-4735-aaac-a26a54acce1f-kube-api-access-hnsq5\") pod \"auto-csr-approver-29562344-6kk6q\" (UID: \"a530431d-cc3a-4735-aaac-a26a54acce1f\") " pod="openshift-infra/auto-csr-approver-29562344-6kk6q" Mar 17 09:44:00 crc kubenswrapper[4813]: I0317 09:44:00.307848 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnsq5\" (UniqueName: \"kubernetes.io/projected/a530431d-cc3a-4735-aaac-a26a54acce1f-kube-api-access-hnsq5\") pod \"auto-csr-approver-29562344-6kk6q\" (UID: \"a530431d-cc3a-4735-aaac-a26a54acce1f\") " pod="openshift-infra/auto-csr-approver-29562344-6kk6q" Mar 17 09:44:00 crc kubenswrapper[4813]: I0317 09:44:00.328818 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnsq5\" (UniqueName: \"kubernetes.io/projected/a530431d-cc3a-4735-aaac-a26a54acce1f-kube-api-access-hnsq5\") pod \"auto-csr-approver-29562344-6kk6q\" (UID: \"a530431d-cc3a-4735-aaac-a26a54acce1f\") " pod="openshift-infra/auto-csr-approver-29562344-6kk6q" Mar 17 09:44:00 crc kubenswrapper[4813]: I0317 09:44:00.471893 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562344-6kk6q" Mar 17 09:44:01 crc kubenswrapper[4813]: I0317 09:44:01.012441 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562344-6kk6q"] Mar 17 09:44:01 crc kubenswrapper[4813]: W0317 09:44:01.023824 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda530431d_cc3a_4735_aaac_a26a54acce1f.slice/crio-91fa8adfd10eec172d255b6230071fb35e76b380154bd3df075cda0953d501f2 WatchSource:0}: Error finding container 91fa8adfd10eec172d255b6230071fb35e76b380154bd3df075cda0953d501f2: Status 404 returned error can't find the container with id 91fa8adfd10eec172d255b6230071fb35e76b380154bd3df075cda0953d501f2 Mar 17 09:44:01 crc kubenswrapper[4813]: I0317 09:44:01.192646 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562344-6kk6q" event={"ID":"a530431d-cc3a-4735-aaac-a26a54acce1f","Type":"ContainerStarted","Data":"91fa8adfd10eec172d255b6230071fb35e76b380154bd3df075cda0953d501f2"} Mar 17 09:44:03 crc kubenswrapper[4813]: I0317 09:44:03.218699 4813 generic.go:334] "Generic (PLEG): container finished" podID="a530431d-cc3a-4735-aaac-a26a54acce1f" containerID="297b8b57c58c954c92e9132a175ce9254e8182b45460d2aebd5aeecf52b0eb7b" exitCode=0 Mar 17 09:44:03 crc kubenswrapper[4813]: I0317 09:44:03.218798 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562344-6kk6q" event={"ID":"a530431d-cc3a-4735-aaac-a26a54acce1f","Type":"ContainerDied","Data":"297b8b57c58c954c92e9132a175ce9254e8182b45460d2aebd5aeecf52b0eb7b"} Mar 17 09:44:04 crc kubenswrapper[4813]: I0317 09:44:04.702232 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562344-6kk6q" Mar 17 09:44:04 crc kubenswrapper[4813]: I0317 09:44:04.830183 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnsq5\" (UniqueName: \"kubernetes.io/projected/a530431d-cc3a-4735-aaac-a26a54acce1f-kube-api-access-hnsq5\") pod \"a530431d-cc3a-4735-aaac-a26a54acce1f\" (UID: \"a530431d-cc3a-4735-aaac-a26a54acce1f\") " Mar 17 09:44:04 crc kubenswrapper[4813]: I0317 09:44:04.840564 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a530431d-cc3a-4735-aaac-a26a54acce1f-kube-api-access-hnsq5" (OuterVolumeSpecName: "kube-api-access-hnsq5") pod "a530431d-cc3a-4735-aaac-a26a54acce1f" (UID: "a530431d-cc3a-4735-aaac-a26a54acce1f"). InnerVolumeSpecName "kube-api-access-hnsq5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:44:04 crc kubenswrapper[4813]: I0317 09:44:04.934488 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnsq5\" (UniqueName: \"kubernetes.io/projected/a530431d-cc3a-4735-aaac-a26a54acce1f-kube-api-access-hnsq5\") on node \"crc\" DevicePath \"\"" Mar 17 09:44:05 crc kubenswrapper[4813]: I0317 09:44:05.248762 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562344-6kk6q" event={"ID":"a530431d-cc3a-4735-aaac-a26a54acce1f","Type":"ContainerDied","Data":"91fa8adfd10eec172d255b6230071fb35e76b380154bd3df075cda0953d501f2"} Mar 17 09:44:05 crc kubenswrapper[4813]: I0317 09:44:05.248838 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91fa8adfd10eec172d255b6230071fb35e76b380154bd3df075cda0953d501f2" Mar 17 09:44:05 crc kubenswrapper[4813]: I0317 09:44:05.248796 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562344-6kk6q" Mar 17 09:44:05 crc kubenswrapper[4813]: I0317 09:44:05.772920 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562338-qjcvb"] Mar 17 09:44:05 crc kubenswrapper[4813]: I0317 09:44:05.780705 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562338-qjcvb"] Mar 17 09:44:06 crc kubenswrapper[4813]: I0317 09:44:06.750586 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ef5582f-769e-40ee-bcf1-56d5509d19e6" path="/var/lib/kubelet/pods/6ef5582f-769e-40ee-bcf1-56d5509d19e6/volumes" Mar 17 09:44:14 crc kubenswrapper[4813]: I0317 09:44:14.113701 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:44:14 crc kubenswrapper[4813]: I0317 09:44:14.114446 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:44:35 crc kubenswrapper[4813]: I0317 09:44:35.587653 4813 generic.go:334] "Generic (PLEG): container finished" podID="d8cecca4-b584-4537-810f-02c1818d2e99" containerID="38b171d4c8baab1dd27791ffc6e04e142a9035567145de55c14c6198535f264f" exitCode=0 Mar 17 09:44:35 crc kubenswrapper[4813]: I0317 09:44:35.587715 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" event={"ID":"d8cecca4-b584-4537-810f-02c1818d2e99","Type":"ContainerDied","Data":"38b171d4c8baab1dd27791ffc6e04e142a9035567145de55c14c6198535f264f"} Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.086352 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.249040 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-ovn-default-certs-0\") pod \"d8cecca4-b584-4537-810f-02c1818d2e99\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.249097 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcz9d\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-kube-api-access-lcz9d\") pod \"d8cecca4-b584-4537-810f-02c1818d2e99\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.249148 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-bootstrap-combined-ca-bundle\") pod \"d8cecca4-b584-4537-810f-02c1818d2e99\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.249169 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-ovn-combined-ca-bundle\") pod \"d8cecca4-b584-4537-810f-02c1818d2e99\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.249202 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-inventory\") pod \"d8cecca4-b584-4537-810f-02c1818d2e99\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.249261 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-ssh-key-openstack-edpm-ipam\") pod \"d8cecca4-b584-4537-810f-02c1818d2e99\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.249298 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"d8cecca4-b584-4537-810f-02c1818d2e99\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.249321 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"d8cecca4-b584-4537-810f-02c1818d2e99\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.249378 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-telemetry-combined-ca-bundle\") pod \"d8cecca4-b584-4537-810f-02c1818d2e99\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.249415 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-nova-combined-ca-bundle\") pod \"d8cecca4-b584-4537-810f-02c1818d2e99\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.249491 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-libvirt-combined-ca-bundle\") pod \"d8cecca4-b584-4537-810f-02c1818d2e99\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.249515 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"d8cecca4-b584-4537-810f-02c1818d2e99\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.249576 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-neutron-metadata-combined-ca-bundle\") pod \"d8cecca4-b584-4537-810f-02c1818d2e99\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.249627 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-repo-setup-combined-ca-bundle\") pod \"d8cecca4-b584-4537-810f-02c1818d2e99\" (UID: \"d8cecca4-b584-4537-810f-02c1818d2e99\") " Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.256726 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "d8cecca4-b584-4537-810f-02c1818d2e99" (UID: "d8cecca4-b584-4537-810f-02c1818d2e99"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.257780 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "d8cecca4-b584-4537-810f-02c1818d2e99" (UID: "d8cecca4-b584-4537-810f-02c1818d2e99"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.257846 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "d8cecca4-b584-4537-810f-02c1818d2e99" (UID: "d8cecca4-b584-4537-810f-02c1818d2e99"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.257937 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "d8cecca4-b584-4537-810f-02c1818d2e99" (UID: "d8cecca4-b584-4537-810f-02c1818d2e99"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.260562 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "d8cecca4-b584-4537-810f-02c1818d2e99" (UID: "d8cecca4-b584-4537-810f-02c1818d2e99"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.260829 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "d8cecca4-b584-4537-810f-02c1818d2e99" (UID: "d8cecca4-b584-4537-810f-02c1818d2e99"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.260845 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "d8cecca4-b584-4537-810f-02c1818d2e99" (UID: "d8cecca4-b584-4537-810f-02c1818d2e99"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.261945 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "d8cecca4-b584-4537-810f-02c1818d2e99" (UID: "d8cecca4-b584-4537-810f-02c1818d2e99"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.262329 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "d8cecca4-b584-4537-810f-02c1818d2e99" (UID: "d8cecca4-b584-4537-810f-02c1818d2e99"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.268346 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "d8cecca4-b584-4537-810f-02c1818d2e99" (UID: "d8cecca4-b584-4537-810f-02c1818d2e99"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.268792 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "d8cecca4-b584-4537-810f-02c1818d2e99" (UID: "d8cecca4-b584-4537-810f-02c1818d2e99"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.269558 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-kube-api-access-lcz9d" (OuterVolumeSpecName: "kube-api-access-lcz9d") pod "d8cecca4-b584-4537-810f-02c1818d2e99" (UID: "d8cecca4-b584-4537-810f-02c1818d2e99"). InnerVolumeSpecName "kube-api-access-lcz9d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.283262 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "d8cecca4-b584-4537-810f-02c1818d2e99" (UID: "d8cecca4-b584-4537-810f-02c1818d2e99"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.300130 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-inventory" (OuterVolumeSpecName: "inventory") pod "d8cecca4-b584-4537-810f-02c1818d2e99" (UID: "d8cecca4-b584-4537-810f-02c1818d2e99"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.353653 4813 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.353726 4813 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.353757 4813 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.353785 4813 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.353812 4813 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.353836 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcz9d\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-kube-api-access-lcz9d\") on node \"crc\" DevicePath \"\"" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.353858 4813 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.353884 4813 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.353907 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.353930 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.353953 4813 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.353980 4813 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/d8cecca4-b584-4537-810f-02c1818d2e99-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.354038 4813 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.354062 4813 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8cecca4-b584-4537-810f-02c1818d2e99-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.613770 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" event={"ID":"d8cecca4-b584-4537-810f-02c1818d2e99","Type":"ContainerDied","Data":"b0eef6fbe7476b2fe451ddd27c0dd9df52d3e0ac26264e67de329bb648d923f8"} Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.613816 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.613837 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b0eef6fbe7476b2fe451ddd27c0dd9df52d3e0ac26264e67de329bb648d923f8" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.725861 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr"] Mar 17 09:44:37 crc kubenswrapper[4813]: E0317 09:44:37.726228 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8cecca4-b584-4537-810f-02c1818d2e99" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.726246 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8cecca4-b584-4537-810f-02c1818d2e99" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 17 09:44:37 crc kubenswrapper[4813]: E0317 09:44:37.726260 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a530431d-cc3a-4735-aaac-a26a54acce1f" containerName="oc" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.726267 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a530431d-cc3a-4735-aaac-a26a54acce1f" containerName="oc" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.726453 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8cecca4-b584-4537-810f-02c1818d2e99" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.726475 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a530431d-cc3a-4735-aaac-a26a54acce1f" containerName="oc" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.727057 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.729147 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.729426 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.729460 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.731206 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hj9h4" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.731236 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.764704 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr"] Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.770702 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mn8bl\" (UniqueName: \"kubernetes.io/projected/5e4d8ae8-8ff6-4969-a349-813dde113094-kube-api-access-mn8bl\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-c5ttr\" (UID: \"5e4d8ae8-8ff6-4969-a349-813dde113094\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.770888 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e4d8ae8-8ff6-4969-a349-813dde113094-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-c5ttr\" (UID: \"5e4d8ae8-8ff6-4969-a349-813dde113094\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.771221 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e4d8ae8-8ff6-4969-a349-813dde113094-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-c5ttr\" (UID: \"5e4d8ae8-8ff6-4969-a349-813dde113094\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.771320 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5e4d8ae8-8ff6-4969-a349-813dde113094-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-c5ttr\" (UID: \"5e4d8ae8-8ff6-4969-a349-813dde113094\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.771803 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e4d8ae8-8ff6-4969-a349-813dde113094-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-c5ttr\" (UID: \"5e4d8ae8-8ff6-4969-a349-813dde113094\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" Mar 17 09:44:37 crc kubenswrapper[4813]: E0317 09:44:37.817137 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8cecca4_b584_4537_810f_02c1818d2e99.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd8cecca4_b584_4537_810f_02c1818d2e99.slice/crio-b0eef6fbe7476b2fe451ddd27c0dd9df52d3e0ac26264e67de329bb648d923f8\": RecentStats: unable to find data in memory cache]" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.875728 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e4d8ae8-8ff6-4969-a349-813dde113094-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-c5ttr\" (UID: \"5e4d8ae8-8ff6-4969-a349-813dde113094\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.875837 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mn8bl\" (UniqueName: \"kubernetes.io/projected/5e4d8ae8-8ff6-4969-a349-813dde113094-kube-api-access-mn8bl\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-c5ttr\" (UID: \"5e4d8ae8-8ff6-4969-a349-813dde113094\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.875861 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e4d8ae8-8ff6-4969-a349-813dde113094-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-c5ttr\" (UID: \"5e4d8ae8-8ff6-4969-a349-813dde113094\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.875908 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e4d8ae8-8ff6-4969-a349-813dde113094-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-c5ttr\" (UID: \"5e4d8ae8-8ff6-4969-a349-813dde113094\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.875937 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5e4d8ae8-8ff6-4969-a349-813dde113094-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-c5ttr\" (UID: \"5e4d8ae8-8ff6-4969-a349-813dde113094\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.877309 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5e4d8ae8-8ff6-4969-a349-813dde113094-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-c5ttr\" (UID: \"5e4d8ae8-8ff6-4969-a349-813dde113094\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.880968 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e4d8ae8-8ff6-4969-a349-813dde113094-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-c5ttr\" (UID: \"5e4d8ae8-8ff6-4969-a349-813dde113094\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.882133 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e4d8ae8-8ff6-4969-a349-813dde113094-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-c5ttr\" (UID: \"5e4d8ae8-8ff6-4969-a349-813dde113094\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.882235 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e4d8ae8-8ff6-4969-a349-813dde113094-ssh-key-openstack-edpm-ipam\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-c5ttr\" (UID: \"5e4d8ae8-8ff6-4969-a349-813dde113094\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" Mar 17 09:44:37 crc kubenswrapper[4813]: I0317 09:44:37.892537 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mn8bl\" (UniqueName: \"kubernetes.io/projected/5e4d8ae8-8ff6-4969-a349-813dde113094-kube-api-access-mn8bl\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-c5ttr\" (UID: \"5e4d8ae8-8ff6-4969-a349-813dde113094\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" Mar 17 09:44:38 crc kubenswrapper[4813]: I0317 09:44:38.055704 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" Mar 17 09:44:38 crc kubenswrapper[4813]: I0317 09:44:38.448104 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr"] Mar 17 09:44:38 crc kubenswrapper[4813]: W0317 09:44:38.453004 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e4d8ae8_8ff6_4969_a349_813dde113094.slice/crio-3562c103ce5e48a543271c2dd8b84759c5af9f9d252ca7c3ca51c5a1971aa447 WatchSource:0}: Error finding container 3562c103ce5e48a543271c2dd8b84759c5af9f9d252ca7c3ca51c5a1971aa447: Status 404 returned error can't find the container with id 3562c103ce5e48a543271c2dd8b84759c5af9f9d252ca7c3ca51c5a1971aa447 Mar 17 09:44:38 crc kubenswrapper[4813]: I0317 09:44:38.627656 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" event={"ID":"5e4d8ae8-8ff6-4969-a349-813dde113094","Type":"ContainerStarted","Data":"3562c103ce5e48a543271c2dd8b84759c5af9f9d252ca7c3ca51c5a1971aa447"} Mar 17 09:44:39 crc kubenswrapper[4813]: I0317 09:44:39.236837 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 09:44:39 crc kubenswrapper[4813]: I0317 09:44:39.642860 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" event={"ID":"5e4d8ae8-8ff6-4969-a349-813dde113094","Type":"ContainerStarted","Data":"2ae66e1640b8b81d8cebb9f7f86db75e3594f81fac08efa97c20dfc87d032ae9"} Mar 17 09:44:39 crc kubenswrapper[4813]: I0317 09:44:39.676843 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" podStartSLOduration=1.899733713 podStartE2EDuration="2.676816655s" podCreationTimestamp="2026-03-17 09:44:37 +0000 UTC" firstStartedPulling="2026-03-17 09:44:38.456242965 +0000 UTC m=+2100.557046474" lastFinishedPulling="2026-03-17 09:44:39.233325887 +0000 UTC m=+2101.334129416" observedRunningTime="2026-03-17 09:44:39.667185665 +0000 UTC m=+2101.767989184" watchObservedRunningTime="2026-03-17 09:44:39.676816655 +0000 UTC m=+2101.777620184" Mar 17 09:44:44 crc kubenswrapper[4813]: I0317 09:44:44.114423 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:44:44 crc kubenswrapper[4813]: I0317 09:44:44.115105 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:44:48 crc kubenswrapper[4813]: I0317 09:44:48.552601 4813 scope.go:117] "RemoveContainer" containerID="39837b8fda3965d754da6e2467aa1025f4b7f30d3903d9901da74fa873d3e1db" Mar 17 09:44:48 crc kubenswrapper[4813]: I0317 09:44:48.586073 4813 scope.go:117] "RemoveContainer" containerID="f129d5928c7aa0b103d13c302017e8fb28643b797037fdeccd6472f22d221548" Mar 17 09:44:58 crc kubenswrapper[4813]: I0317 09:44:58.066484 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4lfn6"] Mar 17 09:44:58 crc kubenswrapper[4813]: I0317 09:44:58.069267 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4lfn6" Mar 17 09:44:58 crc kubenswrapper[4813]: I0317 09:44:58.074011 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4lfn6"] Mar 17 09:44:58 crc kubenswrapper[4813]: I0317 09:44:58.215066 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cadc4299-3256-40ba-b78c-8f863aff4254-catalog-content\") pod \"redhat-operators-4lfn6\" (UID: \"cadc4299-3256-40ba-b78c-8f863aff4254\") " pod="openshift-marketplace/redhat-operators-4lfn6" Mar 17 09:44:58 crc kubenswrapper[4813]: I0317 09:44:58.215134 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bsd86\" (UniqueName: \"kubernetes.io/projected/cadc4299-3256-40ba-b78c-8f863aff4254-kube-api-access-bsd86\") pod \"redhat-operators-4lfn6\" (UID: \"cadc4299-3256-40ba-b78c-8f863aff4254\") " pod="openshift-marketplace/redhat-operators-4lfn6" Mar 17 09:44:58 crc kubenswrapper[4813]: I0317 09:44:58.215175 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cadc4299-3256-40ba-b78c-8f863aff4254-utilities\") pod \"redhat-operators-4lfn6\" (UID: \"cadc4299-3256-40ba-b78c-8f863aff4254\") " pod="openshift-marketplace/redhat-operators-4lfn6" Mar 17 09:44:58 crc kubenswrapper[4813]: I0317 09:44:58.317196 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cadc4299-3256-40ba-b78c-8f863aff4254-catalog-content\") pod \"redhat-operators-4lfn6\" (UID: \"cadc4299-3256-40ba-b78c-8f863aff4254\") " pod="openshift-marketplace/redhat-operators-4lfn6" Mar 17 09:44:58 crc kubenswrapper[4813]: I0317 09:44:58.317263 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bsd86\" (UniqueName: \"kubernetes.io/projected/cadc4299-3256-40ba-b78c-8f863aff4254-kube-api-access-bsd86\") pod \"redhat-operators-4lfn6\" (UID: \"cadc4299-3256-40ba-b78c-8f863aff4254\") " pod="openshift-marketplace/redhat-operators-4lfn6" Mar 17 09:44:58 crc kubenswrapper[4813]: I0317 09:44:58.317300 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cadc4299-3256-40ba-b78c-8f863aff4254-utilities\") pod \"redhat-operators-4lfn6\" (UID: \"cadc4299-3256-40ba-b78c-8f863aff4254\") " pod="openshift-marketplace/redhat-operators-4lfn6" Mar 17 09:44:58 crc kubenswrapper[4813]: I0317 09:44:58.317751 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cadc4299-3256-40ba-b78c-8f863aff4254-catalog-content\") pod \"redhat-operators-4lfn6\" (UID: \"cadc4299-3256-40ba-b78c-8f863aff4254\") " pod="openshift-marketplace/redhat-operators-4lfn6" Mar 17 09:44:58 crc kubenswrapper[4813]: I0317 09:44:58.317797 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cadc4299-3256-40ba-b78c-8f863aff4254-utilities\") pod \"redhat-operators-4lfn6\" (UID: \"cadc4299-3256-40ba-b78c-8f863aff4254\") " pod="openshift-marketplace/redhat-operators-4lfn6" Mar 17 09:44:58 crc kubenswrapper[4813]: I0317 09:44:58.341110 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bsd86\" (UniqueName: \"kubernetes.io/projected/cadc4299-3256-40ba-b78c-8f863aff4254-kube-api-access-bsd86\") pod \"redhat-operators-4lfn6\" (UID: \"cadc4299-3256-40ba-b78c-8f863aff4254\") " pod="openshift-marketplace/redhat-operators-4lfn6" Mar 17 09:44:58 crc kubenswrapper[4813]: I0317 09:44:58.399215 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4lfn6" Mar 17 09:44:58 crc kubenswrapper[4813]: I0317 09:44:58.890496 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4lfn6"] Mar 17 09:44:59 crc kubenswrapper[4813]: I0317 09:44:59.879127 4813 generic.go:334] "Generic (PLEG): container finished" podID="cadc4299-3256-40ba-b78c-8f863aff4254" containerID="62aeb30d754d95983ad4d5ba88debf6ddb52947d5bcbcba0c7905d3a30cc4696" exitCode=0 Mar 17 09:44:59 crc kubenswrapper[4813]: I0317 09:44:59.879220 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4lfn6" event={"ID":"cadc4299-3256-40ba-b78c-8f863aff4254","Type":"ContainerDied","Data":"62aeb30d754d95983ad4d5ba88debf6ddb52947d5bcbcba0c7905d3a30cc4696"} Mar 17 09:44:59 crc kubenswrapper[4813]: I0317 09:44:59.879397 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4lfn6" event={"ID":"cadc4299-3256-40ba-b78c-8f863aff4254","Type":"ContainerStarted","Data":"ef0979dcf3182a7cdf8a751eeda86a6e1f22c282019c1f929ff3994fdff0d593"} Mar 17 09:45:00 crc kubenswrapper[4813]: I0317 09:45:00.162019 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562345-6fwdr"] Mar 17 09:45:00 crc kubenswrapper[4813]: I0317 09:45:00.163343 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562345-6fwdr" Mar 17 09:45:00 crc kubenswrapper[4813]: I0317 09:45:00.165756 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 17 09:45:00 crc kubenswrapper[4813]: I0317 09:45:00.166060 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 17 09:45:00 crc kubenswrapper[4813]: I0317 09:45:00.192356 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562345-6fwdr"] Mar 17 09:45:00 crc kubenswrapper[4813]: I0317 09:45:00.259012 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/037ce2f2-72ef-4a31-9caf-4a2639faace9-config-volume\") pod \"collect-profiles-29562345-6fwdr\" (UID: \"037ce2f2-72ef-4a31-9caf-4a2639faace9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562345-6fwdr" Mar 17 09:45:00 crc kubenswrapper[4813]: I0317 09:45:00.259107 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/037ce2f2-72ef-4a31-9caf-4a2639faace9-secret-volume\") pod \"collect-profiles-29562345-6fwdr\" (UID: \"037ce2f2-72ef-4a31-9caf-4a2639faace9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562345-6fwdr" Mar 17 09:45:00 crc kubenswrapper[4813]: I0317 09:45:00.259343 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pmffw\" (UniqueName: \"kubernetes.io/projected/037ce2f2-72ef-4a31-9caf-4a2639faace9-kube-api-access-pmffw\") pod \"collect-profiles-29562345-6fwdr\" (UID: \"037ce2f2-72ef-4a31-9caf-4a2639faace9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562345-6fwdr" Mar 17 09:45:00 crc kubenswrapper[4813]: I0317 09:45:00.361169 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/037ce2f2-72ef-4a31-9caf-4a2639faace9-config-volume\") pod \"collect-profiles-29562345-6fwdr\" (UID: \"037ce2f2-72ef-4a31-9caf-4a2639faace9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562345-6fwdr" Mar 17 09:45:00 crc kubenswrapper[4813]: I0317 09:45:00.361299 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/037ce2f2-72ef-4a31-9caf-4a2639faace9-secret-volume\") pod \"collect-profiles-29562345-6fwdr\" (UID: \"037ce2f2-72ef-4a31-9caf-4a2639faace9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562345-6fwdr" Mar 17 09:45:00 crc kubenswrapper[4813]: I0317 09:45:00.361392 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pmffw\" (UniqueName: \"kubernetes.io/projected/037ce2f2-72ef-4a31-9caf-4a2639faace9-kube-api-access-pmffw\") pod \"collect-profiles-29562345-6fwdr\" (UID: \"037ce2f2-72ef-4a31-9caf-4a2639faace9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562345-6fwdr" Mar 17 09:45:00 crc kubenswrapper[4813]: I0317 09:45:00.362819 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/037ce2f2-72ef-4a31-9caf-4a2639faace9-config-volume\") pod \"collect-profiles-29562345-6fwdr\" (UID: \"037ce2f2-72ef-4a31-9caf-4a2639faace9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562345-6fwdr" Mar 17 09:45:00 crc kubenswrapper[4813]: I0317 09:45:00.374415 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/037ce2f2-72ef-4a31-9caf-4a2639faace9-secret-volume\") pod \"collect-profiles-29562345-6fwdr\" (UID: \"037ce2f2-72ef-4a31-9caf-4a2639faace9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562345-6fwdr" Mar 17 09:45:00 crc kubenswrapper[4813]: I0317 09:45:00.388343 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pmffw\" (UniqueName: \"kubernetes.io/projected/037ce2f2-72ef-4a31-9caf-4a2639faace9-kube-api-access-pmffw\") pod \"collect-profiles-29562345-6fwdr\" (UID: \"037ce2f2-72ef-4a31-9caf-4a2639faace9\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562345-6fwdr" Mar 17 09:45:00 crc kubenswrapper[4813]: I0317 09:45:00.486113 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562345-6fwdr" Mar 17 09:45:00 crc kubenswrapper[4813]: I0317 09:45:00.808918 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562345-6fwdr"] Mar 17 09:45:00 crc kubenswrapper[4813]: W0317 09:45:00.809670 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod037ce2f2_72ef_4a31_9caf_4a2639faace9.slice/crio-279ec3798e11f5ee202f119294c8f9e2792ff1a30095738ee65e6078a6555db7 WatchSource:0}: Error finding container 279ec3798e11f5ee202f119294c8f9e2792ff1a30095738ee65e6078a6555db7: Status 404 returned error can't find the container with id 279ec3798e11f5ee202f119294c8f9e2792ff1a30095738ee65e6078a6555db7 Mar 17 09:45:00 crc kubenswrapper[4813]: I0317 09:45:00.889432 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4lfn6" event={"ID":"cadc4299-3256-40ba-b78c-8f863aff4254","Type":"ContainerStarted","Data":"9daf775e4e113541bbcc9867c5bc8f68e0cea6d888dff4d90d18cf70ab640028"} Mar 17 09:45:00 crc kubenswrapper[4813]: I0317 09:45:00.890729 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562345-6fwdr" event={"ID":"037ce2f2-72ef-4a31-9caf-4a2639faace9","Type":"ContainerStarted","Data":"279ec3798e11f5ee202f119294c8f9e2792ff1a30095738ee65e6078a6555db7"} Mar 17 09:45:01 crc kubenswrapper[4813]: I0317 09:45:01.904620 4813 generic.go:334] "Generic (PLEG): container finished" podID="037ce2f2-72ef-4a31-9caf-4a2639faace9" containerID="134661704d56f18804c48ed4821ef7d4120c3f89af802a4e4cdc6d88ea98c44c" exitCode=0 Mar 17 09:45:01 crc kubenswrapper[4813]: I0317 09:45:01.904678 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562345-6fwdr" event={"ID":"037ce2f2-72ef-4a31-9caf-4a2639faace9","Type":"ContainerDied","Data":"134661704d56f18804c48ed4821ef7d4120c3f89af802a4e4cdc6d88ea98c44c"} Mar 17 09:45:03 crc kubenswrapper[4813]: I0317 09:45:03.378646 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562345-6fwdr" Mar 17 09:45:03 crc kubenswrapper[4813]: I0317 09:45:03.528402 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/037ce2f2-72ef-4a31-9caf-4a2639faace9-secret-volume\") pod \"037ce2f2-72ef-4a31-9caf-4a2639faace9\" (UID: \"037ce2f2-72ef-4a31-9caf-4a2639faace9\") " Mar 17 09:45:03 crc kubenswrapper[4813]: I0317 09:45:03.528775 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pmffw\" (UniqueName: \"kubernetes.io/projected/037ce2f2-72ef-4a31-9caf-4a2639faace9-kube-api-access-pmffw\") pod \"037ce2f2-72ef-4a31-9caf-4a2639faace9\" (UID: \"037ce2f2-72ef-4a31-9caf-4a2639faace9\") " Mar 17 09:45:03 crc kubenswrapper[4813]: I0317 09:45:03.528873 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/037ce2f2-72ef-4a31-9caf-4a2639faace9-config-volume\") pod \"037ce2f2-72ef-4a31-9caf-4a2639faace9\" (UID: \"037ce2f2-72ef-4a31-9caf-4a2639faace9\") " Mar 17 09:45:03 crc kubenswrapper[4813]: I0317 09:45:03.529588 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/037ce2f2-72ef-4a31-9caf-4a2639faace9-config-volume" (OuterVolumeSpecName: "config-volume") pod "037ce2f2-72ef-4a31-9caf-4a2639faace9" (UID: "037ce2f2-72ef-4a31-9caf-4a2639faace9"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:45:03 crc kubenswrapper[4813]: I0317 09:45:03.536293 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/037ce2f2-72ef-4a31-9caf-4a2639faace9-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "037ce2f2-72ef-4a31-9caf-4a2639faace9" (UID: "037ce2f2-72ef-4a31-9caf-4a2639faace9"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:45:03 crc kubenswrapper[4813]: I0317 09:45:03.536665 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/037ce2f2-72ef-4a31-9caf-4a2639faace9-kube-api-access-pmffw" (OuterVolumeSpecName: "kube-api-access-pmffw") pod "037ce2f2-72ef-4a31-9caf-4a2639faace9" (UID: "037ce2f2-72ef-4a31-9caf-4a2639faace9"). InnerVolumeSpecName "kube-api-access-pmffw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:45:03 crc kubenswrapper[4813]: I0317 09:45:03.631522 4813 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/037ce2f2-72ef-4a31-9caf-4a2639faace9-config-volume\") on node \"crc\" DevicePath \"\"" Mar 17 09:45:03 crc kubenswrapper[4813]: I0317 09:45:03.631555 4813 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/037ce2f2-72ef-4a31-9caf-4a2639faace9-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 17 09:45:03 crc kubenswrapper[4813]: I0317 09:45:03.631565 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pmffw\" (UniqueName: \"kubernetes.io/projected/037ce2f2-72ef-4a31-9caf-4a2639faace9-kube-api-access-pmffw\") on node \"crc\" DevicePath \"\"" Mar 17 09:45:03 crc kubenswrapper[4813]: I0317 09:45:03.926913 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562345-6fwdr" event={"ID":"037ce2f2-72ef-4a31-9caf-4a2639faace9","Type":"ContainerDied","Data":"279ec3798e11f5ee202f119294c8f9e2792ff1a30095738ee65e6078a6555db7"} Mar 17 09:45:03 crc kubenswrapper[4813]: I0317 09:45:03.927267 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="279ec3798e11f5ee202f119294c8f9e2792ff1a30095738ee65e6078a6555db7" Mar 17 09:45:03 crc kubenswrapper[4813]: I0317 09:45:03.927168 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562345-6fwdr" Mar 17 09:45:04 crc kubenswrapper[4813]: I0317 09:45:04.462849 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7"] Mar 17 09:45:04 crc kubenswrapper[4813]: I0317 09:45:04.476677 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562300-vntp7"] Mar 17 09:45:04 crc kubenswrapper[4813]: I0317 09:45:04.745985 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b3992349-bc61-405e-be75-609cce2fad10" path="/var/lib/kubelet/pods/b3992349-bc61-405e-be75-609cce2fad10/volumes" Mar 17 09:45:04 crc kubenswrapper[4813]: I0317 09:45:04.941653 4813 generic.go:334] "Generic (PLEG): container finished" podID="cadc4299-3256-40ba-b78c-8f863aff4254" containerID="9daf775e4e113541bbcc9867c5bc8f68e0cea6d888dff4d90d18cf70ab640028" exitCode=0 Mar 17 09:45:04 crc kubenswrapper[4813]: I0317 09:45:04.941726 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4lfn6" event={"ID":"cadc4299-3256-40ba-b78c-8f863aff4254","Type":"ContainerDied","Data":"9daf775e4e113541bbcc9867c5bc8f68e0cea6d888dff4d90d18cf70ab640028"} Mar 17 09:45:05 crc kubenswrapper[4813]: I0317 09:45:05.955632 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4lfn6" event={"ID":"cadc4299-3256-40ba-b78c-8f863aff4254","Type":"ContainerStarted","Data":"0832b0a87d5d61938e992257aa4171f1425e6c93c018eb2100c0df3fd2a333f4"} Mar 17 09:45:05 crc kubenswrapper[4813]: I0317 09:45:05.990658 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4lfn6" podStartSLOduration=3.394136975 podStartE2EDuration="8.990639235s" podCreationTimestamp="2026-03-17 09:44:57 +0000 UTC" firstStartedPulling="2026-03-17 09:44:59.88087283 +0000 UTC m=+2121.981676329" lastFinishedPulling="2026-03-17 09:45:05.47737505 +0000 UTC m=+2127.578178589" observedRunningTime="2026-03-17 09:45:05.980627523 +0000 UTC m=+2128.081431022" watchObservedRunningTime="2026-03-17 09:45:05.990639235 +0000 UTC m=+2128.091442734" Mar 17 09:45:08 crc kubenswrapper[4813]: I0317 09:45:08.399525 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4lfn6" Mar 17 09:45:08 crc kubenswrapper[4813]: I0317 09:45:08.399909 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4lfn6" Mar 17 09:45:09 crc kubenswrapper[4813]: I0317 09:45:09.484275 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4lfn6" podUID="cadc4299-3256-40ba-b78c-8f863aff4254" containerName="registry-server" probeResult="failure" output=< Mar 17 09:45:09 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Mar 17 09:45:09 crc kubenswrapper[4813]: > Mar 17 09:45:13 crc kubenswrapper[4813]: I0317 09:45:13.942863 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-846vg"] Mar 17 09:45:13 crc kubenswrapper[4813]: E0317 09:45:13.945545 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="037ce2f2-72ef-4a31-9caf-4a2639faace9" containerName="collect-profiles" Mar 17 09:45:13 crc kubenswrapper[4813]: I0317 09:45:13.945634 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="037ce2f2-72ef-4a31-9caf-4a2639faace9" containerName="collect-profiles" Mar 17 09:45:13 crc kubenswrapper[4813]: I0317 09:45:13.946085 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="037ce2f2-72ef-4a31-9caf-4a2639faace9" containerName="collect-profiles" Mar 17 09:45:13 crc kubenswrapper[4813]: I0317 09:45:13.949389 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-846vg" Mar 17 09:45:13 crc kubenswrapper[4813]: I0317 09:45:13.977282 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-846vg"] Mar 17 09:45:14 crc kubenswrapper[4813]: I0317 09:45:14.037958 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkdzz\" (UniqueName: \"kubernetes.io/projected/f5541565-db4d-44af-9617-952d8077e9f1-kube-api-access-mkdzz\") pod \"community-operators-846vg\" (UID: \"f5541565-db4d-44af-9617-952d8077e9f1\") " pod="openshift-marketplace/community-operators-846vg" Mar 17 09:45:14 crc kubenswrapper[4813]: I0317 09:45:14.038044 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5541565-db4d-44af-9617-952d8077e9f1-catalog-content\") pod \"community-operators-846vg\" (UID: \"f5541565-db4d-44af-9617-952d8077e9f1\") " pod="openshift-marketplace/community-operators-846vg" Mar 17 09:45:14 crc kubenswrapper[4813]: I0317 09:45:14.038406 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5541565-db4d-44af-9617-952d8077e9f1-utilities\") pod \"community-operators-846vg\" (UID: \"f5541565-db4d-44af-9617-952d8077e9f1\") " pod="openshift-marketplace/community-operators-846vg" Mar 17 09:45:14 crc kubenswrapper[4813]: I0317 09:45:14.113990 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:45:14 crc kubenswrapper[4813]: I0317 09:45:14.114067 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:45:14 crc kubenswrapper[4813]: I0317 09:45:14.114123 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:45:14 crc kubenswrapper[4813]: I0317 09:45:14.115091 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"886e3a29e582cbf7efd8021b4dad6057b4824ee4ecc8a4ac9ae12df4687c6449"} pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 09:45:14 crc kubenswrapper[4813]: I0317 09:45:14.115174 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" containerID="cri-o://886e3a29e582cbf7efd8021b4dad6057b4824ee4ecc8a4ac9ae12df4687c6449" gracePeriod=600 Mar 17 09:45:14 crc kubenswrapper[4813]: I0317 09:45:14.140948 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5541565-db4d-44af-9617-952d8077e9f1-utilities\") pod \"community-operators-846vg\" (UID: \"f5541565-db4d-44af-9617-952d8077e9f1\") " pod="openshift-marketplace/community-operators-846vg" Mar 17 09:45:14 crc kubenswrapper[4813]: I0317 09:45:14.141057 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkdzz\" (UniqueName: \"kubernetes.io/projected/f5541565-db4d-44af-9617-952d8077e9f1-kube-api-access-mkdzz\") pod \"community-operators-846vg\" (UID: \"f5541565-db4d-44af-9617-952d8077e9f1\") " pod="openshift-marketplace/community-operators-846vg" Mar 17 09:45:14 crc kubenswrapper[4813]: I0317 09:45:14.141104 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5541565-db4d-44af-9617-952d8077e9f1-catalog-content\") pod \"community-operators-846vg\" (UID: \"f5541565-db4d-44af-9617-952d8077e9f1\") " pod="openshift-marketplace/community-operators-846vg" Mar 17 09:45:14 crc kubenswrapper[4813]: I0317 09:45:14.141692 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5541565-db4d-44af-9617-952d8077e9f1-catalog-content\") pod \"community-operators-846vg\" (UID: \"f5541565-db4d-44af-9617-952d8077e9f1\") " pod="openshift-marketplace/community-operators-846vg" Mar 17 09:45:14 crc kubenswrapper[4813]: I0317 09:45:14.141715 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5541565-db4d-44af-9617-952d8077e9f1-utilities\") pod \"community-operators-846vg\" (UID: \"f5541565-db4d-44af-9617-952d8077e9f1\") " pod="openshift-marketplace/community-operators-846vg" Mar 17 09:45:14 crc kubenswrapper[4813]: I0317 09:45:14.166525 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkdzz\" (UniqueName: \"kubernetes.io/projected/f5541565-db4d-44af-9617-952d8077e9f1-kube-api-access-mkdzz\") pod \"community-operators-846vg\" (UID: \"f5541565-db4d-44af-9617-952d8077e9f1\") " pod="openshift-marketplace/community-operators-846vg" Mar 17 09:45:14 crc kubenswrapper[4813]: I0317 09:45:14.291291 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-846vg" Mar 17 09:45:14 crc kubenswrapper[4813]: I0317 09:45:14.813118 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-846vg"] Mar 17 09:45:15 crc kubenswrapper[4813]: I0317 09:45:15.038538 4813 generic.go:334] "Generic (PLEG): container finished" podID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerID="886e3a29e582cbf7efd8021b4dad6057b4824ee4ecc8a4ac9ae12df4687c6449" exitCode=0 Mar 17 09:45:15 crc kubenswrapper[4813]: I0317 09:45:15.038577 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerDied","Data":"886e3a29e582cbf7efd8021b4dad6057b4824ee4ecc8a4ac9ae12df4687c6449"} Mar 17 09:45:15 crc kubenswrapper[4813]: I0317 09:45:15.038892 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerStarted","Data":"5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4"} Mar 17 09:45:15 crc kubenswrapper[4813]: I0317 09:45:15.038910 4813 scope.go:117] "RemoveContainer" containerID="cb1a56ea5f9d57bf61c57781f46bda3fa54a009c451a1d17cdd4ff0b4674f368" Mar 17 09:45:15 crc kubenswrapper[4813]: I0317 09:45:15.043090 4813 generic.go:334] "Generic (PLEG): container finished" podID="f5541565-db4d-44af-9617-952d8077e9f1" containerID="e7aa8f07dd481c1734a9c89952ea1474e0e1af62be218178ca0df348a8b9ea6f" exitCode=0 Mar 17 09:45:15 crc kubenswrapper[4813]: I0317 09:45:15.043141 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-846vg" event={"ID":"f5541565-db4d-44af-9617-952d8077e9f1","Type":"ContainerDied","Data":"e7aa8f07dd481c1734a9c89952ea1474e0e1af62be218178ca0df348a8b9ea6f"} Mar 17 09:45:15 crc kubenswrapper[4813]: I0317 09:45:15.043175 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-846vg" event={"ID":"f5541565-db4d-44af-9617-952d8077e9f1","Type":"ContainerStarted","Data":"3d0bed98f5ec1baa0384d80bc2c1b7fc38a73a1dae9a3e4672512a0ce5d4a399"} Mar 17 09:45:16 crc kubenswrapper[4813]: I0317 09:45:16.060071 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-846vg" event={"ID":"f5541565-db4d-44af-9617-952d8077e9f1","Type":"ContainerStarted","Data":"ef70af27c21116b384222ecee4e4402fc01c35ed8836851975f667435e53abad"} Mar 17 09:45:17 crc kubenswrapper[4813]: I0317 09:45:17.070915 4813 generic.go:334] "Generic (PLEG): container finished" podID="f5541565-db4d-44af-9617-952d8077e9f1" containerID="ef70af27c21116b384222ecee4e4402fc01c35ed8836851975f667435e53abad" exitCode=0 Mar 17 09:45:17 crc kubenswrapper[4813]: I0317 09:45:17.070960 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-846vg" event={"ID":"f5541565-db4d-44af-9617-952d8077e9f1","Type":"ContainerDied","Data":"ef70af27c21116b384222ecee4e4402fc01c35ed8836851975f667435e53abad"} Mar 17 09:45:18 crc kubenswrapper[4813]: I0317 09:45:18.086373 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-846vg" event={"ID":"f5541565-db4d-44af-9617-952d8077e9f1","Type":"ContainerStarted","Data":"98bceec31818838ad2f50ece237759625c24a89bc971eb4b8980554e57d6bb7a"} Mar 17 09:45:18 crc kubenswrapper[4813]: I0317 09:45:18.113967 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-846vg" podStartSLOduration=2.573773648 podStartE2EDuration="5.113945046s" podCreationTimestamp="2026-03-17 09:45:13 +0000 UTC" firstStartedPulling="2026-03-17 09:45:15.044416411 +0000 UTC m=+2137.145219910" lastFinishedPulling="2026-03-17 09:45:17.584587819 +0000 UTC m=+2139.685391308" observedRunningTime="2026-03-17 09:45:18.111024555 +0000 UTC m=+2140.211828064" watchObservedRunningTime="2026-03-17 09:45:18.113945046 +0000 UTC m=+2140.214748555" Mar 17 09:45:18 crc kubenswrapper[4813]: I0317 09:45:18.479179 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4lfn6" Mar 17 09:45:18 crc kubenswrapper[4813]: I0317 09:45:18.575147 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4lfn6" Mar 17 09:45:20 crc kubenswrapper[4813]: I0317 09:45:20.303401 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4lfn6"] Mar 17 09:45:20 crc kubenswrapper[4813]: I0317 09:45:20.304389 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4lfn6" podUID="cadc4299-3256-40ba-b78c-8f863aff4254" containerName="registry-server" containerID="cri-o://0832b0a87d5d61938e992257aa4171f1425e6c93c018eb2100c0df3fd2a333f4" gracePeriod=2 Mar 17 09:45:20 crc kubenswrapper[4813]: I0317 09:45:20.892090 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4lfn6" Mar 17 09:45:20 crc kubenswrapper[4813]: I0317 09:45:20.971981 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bsd86\" (UniqueName: \"kubernetes.io/projected/cadc4299-3256-40ba-b78c-8f863aff4254-kube-api-access-bsd86\") pod \"cadc4299-3256-40ba-b78c-8f863aff4254\" (UID: \"cadc4299-3256-40ba-b78c-8f863aff4254\") " Mar 17 09:45:20 crc kubenswrapper[4813]: I0317 09:45:20.972103 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cadc4299-3256-40ba-b78c-8f863aff4254-catalog-content\") pod \"cadc4299-3256-40ba-b78c-8f863aff4254\" (UID: \"cadc4299-3256-40ba-b78c-8f863aff4254\") " Mar 17 09:45:20 crc kubenswrapper[4813]: I0317 09:45:20.972194 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cadc4299-3256-40ba-b78c-8f863aff4254-utilities\") pod \"cadc4299-3256-40ba-b78c-8f863aff4254\" (UID: \"cadc4299-3256-40ba-b78c-8f863aff4254\") " Mar 17 09:45:20 crc kubenswrapper[4813]: I0317 09:45:20.973306 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cadc4299-3256-40ba-b78c-8f863aff4254-utilities" (OuterVolumeSpecName: "utilities") pod "cadc4299-3256-40ba-b78c-8f863aff4254" (UID: "cadc4299-3256-40ba-b78c-8f863aff4254"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:45:20 crc kubenswrapper[4813]: I0317 09:45:20.980764 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cadc4299-3256-40ba-b78c-8f863aff4254-kube-api-access-bsd86" (OuterVolumeSpecName: "kube-api-access-bsd86") pod "cadc4299-3256-40ba-b78c-8f863aff4254" (UID: "cadc4299-3256-40ba-b78c-8f863aff4254"). InnerVolumeSpecName "kube-api-access-bsd86". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:45:21 crc kubenswrapper[4813]: I0317 09:45:21.074747 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bsd86\" (UniqueName: \"kubernetes.io/projected/cadc4299-3256-40ba-b78c-8f863aff4254-kube-api-access-bsd86\") on node \"crc\" DevicePath \"\"" Mar 17 09:45:21 crc kubenswrapper[4813]: I0317 09:45:21.074794 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cadc4299-3256-40ba-b78c-8f863aff4254-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:45:21 crc kubenswrapper[4813]: I0317 09:45:21.099098 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cadc4299-3256-40ba-b78c-8f863aff4254-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cadc4299-3256-40ba-b78c-8f863aff4254" (UID: "cadc4299-3256-40ba-b78c-8f863aff4254"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:45:21 crc kubenswrapper[4813]: I0317 09:45:21.117557 4813 generic.go:334] "Generic (PLEG): container finished" podID="cadc4299-3256-40ba-b78c-8f863aff4254" containerID="0832b0a87d5d61938e992257aa4171f1425e6c93c018eb2100c0df3fd2a333f4" exitCode=0 Mar 17 09:45:21 crc kubenswrapper[4813]: I0317 09:45:21.117668 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4lfn6" Mar 17 09:45:21 crc kubenswrapper[4813]: I0317 09:45:21.117692 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4lfn6" event={"ID":"cadc4299-3256-40ba-b78c-8f863aff4254","Type":"ContainerDied","Data":"0832b0a87d5d61938e992257aa4171f1425e6c93c018eb2100c0df3fd2a333f4"} Mar 17 09:45:21 crc kubenswrapper[4813]: I0317 09:45:21.117790 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4lfn6" event={"ID":"cadc4299-3256-40ba-b78c-8f863aff4254","Type":"ContainerDied","Data":"ef0979dcf3182a7cdf8a751eeda86a6e1f22c282019c1f929ff3994fdff0d593"} Mar 17 09:45:21 crc kubenswrapper[4813]: I0317 09:45:21.117861 4813 scope.go:117] "RemoveContainer" containerID="0832b0a87d5d61938e992257aa4171f1425e6c93c018eb2100c0df3fd2a333f4" Mar 17 09:45:21 crc kubenswrapper[4813]: I0317 09:45:21.158437 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4lfn6"] Mar 17 09:45:21 crc kubenswrapper[4813]: I0317 09:45:21.164658 4813 scope.go:117] "RemoveContainer" containerID="9daf775e4e113541bbcc9867c5bc8f68e0cea6d888dff4d90d18cf70ab640028" Mar 17 09:45:21 crc kubenswrapper[4813]: I0317 09:45:21.165413 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4lfn6"] Mar 17 09:45:21 crc kubenswrapper[4813]: I0317 09:45:21.176873 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cadc4299-3256-40ba-b78c-8f863aff4254-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:45:21 crc kubenswrapper[4813]: I0317 09:45:21.209853 4813 scope.go:117] "RemoveContainer" containerID="62aeb30d754d95983ad4d5ba88debf6ddb52947d5bcbcba0c7905d3a30cc4696" Mar 17 09:45:21 crc kubenswrapper[4813]: I0317 09:45:21.243699 4813 scope.go:117] "RemoveContainer" containerID="0832b0a87d5d61938e992257aa4171f1425e6c93c018eb2100c0df3fd2a333f4" Mar 17 09:45:21 crc kubenswrapper[4813]: E0317 09:45:21.244157 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0832b0a87d5d61938e992257aa4171f1425e6c93c018eb2100c0df3fd2a333f4\": container with ID starting with 0832b0a87d5d61938e992257aa4171f1425e6c93c018eb2100c0df3fd2a333f4 not found: ID does not exist" containerID="0832b0a87d5d61938e992257aa4171f1425e6c93c018eb2100c0df3fd2a333f4" Mar 17 09:45:21 crc kubenswrapper[4813]: I0317 09:45:21.244187 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0832b0a87d5d61938e992257aa4171f1425e6c93c018eb2100c0df3fd2a333f4"} err="failed to get container status \"0832b0a87d5d61938e992257aa4171f1425e6c93c018eb2100c0df3fd2a333f4\": rpc error: code = NotFound desc = could not find container \"0832b0a87d5d61938e992257aa4171f1425e6c93c018eb2100c0df3fd2a333f4\": container with ID starting with 0832b0a87d5d61938e992257aa4171f1425e6c93c018eb2100c0df3fd2a333f4 not found: ID does not exist" Mar 17 09:45:21 crc kubenswrapper[4813]: I0317 09:45:21.244218 4813 scope.go:117] "RemoveContainer" containerID="9daf775e4e113541bbcc9867c5bc8f68e0cea6d888dff4d90d18cf70ab640028" Mar 17 09:45:21 crc kubenswrapper[4813]: E0317 09:45:21.244535 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9daf775e4e113541bbcc9867c5bc8f68e0cea6d888dff4d90d18cf70ab640028\": container with ID starting with 9daf775e4e113541bbcc9867c5bc8f68e0cea6d888dff4d90d18cf70ab640028 not found: ID does not exist" containerID="9daf775e4e113541bbcc9867c5bc8f68e0cea6d888dff4d90d18cf70ab640028" Mar 17 09:45:21 crc kubenswrapper[4813]: I0317 09:45:21.244555 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9daf775e4e113541bbcc9867c5bc8f68e0cea6d888dff4d90d18cf70ab640028"} err="failed to get container status \"9daf775e4e113541bbcc9867c5bc8f68e0cea6d888dff4d90d18cf70ab640028\": rpc error: code = NotFound desc = could not find container \"9daf775e4e113541bbcc9867c5bc8f68e0cea6d888dff4d90d18cf70ab640028\": container with ID starting with 9daf775e4e113541bbcc9867c5bc8f68e0cea6d888dff4d90d18cf70ab640028 not found: ID does not exist" Mar 17 09:45:21 crc kubenswrapper[4813]: I0317 09:45:21.244568 4813 scope.go:117] "RemoveContainer" containerID="62aeb30d754d95983ad4d5ba88debf6ddb52947d5bcbcba0c7905d3a30cc4696" Mar 17 09:45:21 crc kubenswrapper[4813]: E0317 09:45:21.244943 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62aeb30d754d95983ad4d5ba88debf6ddb52947d5bcbcba0c7905d3a30cc4696\": container with ID starting with 62aeb30d754d95983ad4d5ba88debf6ddb52947d5bcbcba0c7905d3a30cc4696 not found: ID does not exist" containerID="62aeb30d754d95983ad4d5ba88debf6ddb52947d5bcbcba0c7905d3a30cc4696" Mar 17 09:45:21 crc kubenswrapper[4813]: I0317 09:45:21.244961 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62aeb30d754d95983ad4d5ba88debf6ddb52947d5bcbcba0c7905d3a30cc4696"} err="failed to get container status \"62aeb30d754d95983ad4d5ba88debf6ddb52947d5bcbcba0c7905d3a30cc4696\": rpc error: code = NotFound desc = could not find container \"62aeb30d754d95983ad4d5ba88debf6ddb52947d5bcbcba0c7905d3a30cc4696\": container with ID starting with 62aeb30d754d95983ad4d5ba88debf6ddb52947d5bcbcba0c7905d3a30cc4696 not found: ID does not exist" Mar 17 09:45:22 crc kubenswrapper[4813]: I0317 09:45:22.750514 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cadc4299-3256-40ba-b78c-8f863aff4254" path="/var/lib/kubelet/pods/cadc4299-3256-40ba-b78c-8f863aff4254/volumes" Mar 17 09:45:24 crc kubenswrapper[4813]: I0317 09:45:24.292178 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-846vg" Mar 17 09:45:24 crc kubenswrapper[4813]: I0317 09:45:24.293690 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-846vg" Mar 17 09:45:24 crc kubenswrapper[4813]: I0317 09:45:24.361160 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-846vg" Mar 17 09:45:25 crc kubenswrapper[4813]: I0317 09:45:25.234442 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-846vg" Mar 17 09:45:25 crc kubenswrapper[4813]: I0317 09:45:25.294675 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-846vg"] Mar 17 09:45:27 crc kubenswrapper[4813]: I0317 09:45:27.189834 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-846vg" podUID="f5541565-db4d-44af-9617-952d8077e9f1" containerName="registry-server" containerID="cri-o://98bceec31818838ad2f50ece237759625c24a89bc971eb4b8980554e57d6bb7a" gracePeriod=2 Mar 17 09:45:27 crc kubenswrapper[4813]: I0317 09:45:27.664437 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-846vg" Mar 17 09:45:27 crc kubenswrapper[4813]: I0317 09:45:27.758659 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mkdzz\" (UniqueName: \"kubernetes.io/projected/f5541565-db4d-44af-9617-952d8077e9f1-kube-api-access-mkdzz\") pod \"f5541565-db4d-44af-9617-952d8077e9f1\" (UID: \"f5541565-db4d-44af-9617-952d8077e9f1\") " Mar 17 09:45:27 crc kubenswrapper[4813]: I0317 09:45:27.758848 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5541565-db4d-44af-9617-952d8077e9f1-utilities\") pod \"f5541565-db4d-44af-9617-952d8077e9f1\" (UID: \"f5541565-db4d-44af-9617-952d8077e9f1\") " Mar 17 09:45:27 crc kubenswrapper[4813]: I0317 09:45:27.758876 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5541565-db4d-44af-9617-952d8077e9f1-catalog-content\") pod \"f5541565-db4d-44af-9617-952d8077e9f1\" (UID: \"f5541565-db4d-44af-9617-952d8077e9f1\") " Mar 17 09:45:27 crc kubenswrapper[4813]: I0317 09:45:27.759723 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5541565-db4d-44af-9617-952d8077e9f1-utilities" (OuterVolumeSpecName: "utilities") pod "f5541565-db4d-44af-9617-952d8077e9f1" (UID: "f5541565-db4d-44af-9617-952d8077e9f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:45:27 crc kubenswrapper[4813]: I0317 09:45:27.764473 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5541565-db4d-44af-9617-952d8077e9f1-kube-api-access-mkdzz" (OuterVolumeSpecName: "kube-api-access-mkdzz") pod "f5541565-db4d-44af-9617-952d8077e9f1" (UID: "f5541565-db4d-44af-9617-952d8077e9f1"). InnerVolumeSpecName "kube-api-access-mkdzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:45:27 crc kubenswrapper[4813]: I0317 09:45:27.829802 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5541565-db4d-44af-9617-952d8077e9f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5541565-db4d-44af-9617-952d8077e9f1" (UID: "f5541565-db4d-44af-9617-952d8077e9f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:45:27 crc kubenswrapper[4813]: I0317 09:45:27.861814 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mkdzz\" (UniqueName: \"kubernetes.io/projected/f5541565-db4d-44af-9617-952d8077e9f1-kube-api-access-mkdzz\") on node \"crc\" DevicePath \"\"" Mar 17 09:45:27 crc kubenswrapper[4813]: I0317 09:45:27.861868 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5541565-db4d-44af-9617-952d8077e9f1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:45:27 crc kubenswrapper[4813]: I0317 09:45:27.861881 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5541565-db4d-44af-9617-952d8077e9f1-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:45:28 crc kubenswrapper[4813]: I0317 09:45:28.199361 4813 generic.go:334] "Generic (PLEG): container finished" podID="f5541565-db4d-44af-9617-952d8077e9f1" containerID="98bceec31818838ad2f50ece237759625c24a89bc971eb4b8980554e57d6bb7a" exitCode=0 Mar 17 09:45:28 crc kubenswrapper[4813]: I0317 09:45:28.199412 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-846vg" event={"ID":"f5541565-db4d-44af-9617-952d8077e9f1","Type":"ContainerDied","Data":"98bceec31818838ad2f50ece237759625c24a89bc971eb4b8980554e57d6bb7a"} Mar 17 09:45:28 crc kubenswrapper[4813]: I0317 09:45:28.199438 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-846vg" event={"ID":"f5541565-db4d-44af-9617-952d8077e9f1","Type":"ContainerDied","Data":"3d0bed98f5ec1baa0384d80bc2c1b7fc38a73a1dae9a3e4672512a0ce5d4a399"} Mar 17 09:45:28 crc kubenswrapper[4813]: I0317 09:45:28.199456 4813 scope.go:117] "RemoveContainer" containerID="98bceec31818838ad2f50ece237759625c24a89bc971eb4b8980554e57d6bb7a" Mar 17 09:45:28 crc kubenswrapper[4813]: I0317 09:45:28.200078 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-846vg" Mar 17 09:45:28 crc kubenswrapper[4813]: I0317 09:45:28.263654 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-846vg"] Mar 17 09:45:28 crc kubenswrapper[4813]: I0317 09:45:28.265308 4813 scope.go:117] "RemoveContainer" containerID="ef70af27c21116b384222ecee4e4402fc01c35ed8836851975f667435e53abad" Mar 17 09:45:28 crc kubenswrapper[4813]: I0317 09:45:28.271772 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-846vg"] Mar 17 09:45:28 crc kubenswrapper[4813]: I0317 09:45:28.289328 4813 scope.go:117] "RemoveContainer" containerID="e7aa8f07dd481c1734a9c89952ea1474e0e1af62be218178ca0df348a8b9ea6f" Mar 17 09:45:28 crc kubenswrapper[4813]: I0317 09:45:28.340023 4813 scope.go:117] "RemoveContainer" containerID="98bceec31818838ad2f50ece237759625c24a89bc971eb4b8980554e57d6bb7a" Mar 17 09:45:28 crc kubenswrapper[4813]: E0317 09:45:28.340661 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98bceec31818838ad2f50ece237759625c24a89bc971eb4b8980554e57d6bb7a\": container with ID starting with 98bceec31818838ad2f50ece237759625c24a89bc971eb4b8980554e57d6bb7a not found: ID does not exist" containerID="98bceec31818838ad2f50ece237759625c24a89bc971eb4b8980554e57d6bb7a" Mar 17 09:45:28 crc kubenswrapper[4813]: I0317 09:45:28.340703 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98bceec31818838ad2f50ece237759625c24a89bc971eb4b8980554e57d6bb7a"} err="failed to get container status \"98bceec31818838ad2f50ece237759625c24a89bc971eb4b8980554e57d6bb7a\": rpc error: code = NotFound desc = could not find container \"98bceec31818838ad2f50ece237759625c24a89bc971eb4b8980554e57d6bb7a\": container with ID starting with 98bceec31818838ad2f50ece237759625c24a89bc971eb4b8980554e57d6bb7a not found: ID does not exist" Mar 17 09:45:28 crc kubenswrapper[4813]: I0317 09:45:28.340730 4813 scope.go:117] "RemoveContainer" containerID="ef70af27c21116b384222ecee4e4402fc01c35ed8836851975f667435e53abad" Mar 17 09:45:28 crc kubenswrapper[4813]: E0317 09:45:28.341024 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef70af27c21116b384222ecee4e4402fc01c35ed8836851975f667435e53abad\": container with ID starting with ef70af27c21116b384222ecee4e4402fc01c35ed8836851975f667435e53abad not found: ID does not exist" containerID="ef70af27c21116b384222ecee4e4402fc01c35ed8836851975f667435e53abad" Mar 17 09:45:28 crc kubenswrapper[4813]: I0317 09:45:28.341046 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef70af27c21116b384222ecee4e4402fc01c35ed8836851975f667435e53abad"} err="failed to get container status \"ef70af27c21116b384222ecee4e4402fc01c35ed8836851975f667435e53abad\": rpc error: code = NotFound desc = could not find container \"ef70af27c21116b384222ecee4e4402fc01c35ed8836851975f667435e53abad\": container with ID starting with ef70af27c21116b384222ecee4e4402fc01c35ed8836851975f667435e53abad not found: ID does not exist" Mar 17 09:45:28 crc kubenswrapper[4813]: I0317 09:45:28.341061 4813 scope.go:117] "RemoveContainer" containerID="e7aa8f07dd481c1734a9c89952ea1474e0e1af62be218178ca0df348a8b9ea6f" Mar 17 09:45:28 crc kubenswrapper[4813]: E0317 09:45:28.341319 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7aa8f07dd481c1734a9c89952ea1474e0e1af62be218178ca0df348a8b9ea6f\": container with ID starting with e7aa8f07dd481c1734a9c89952ea1474e0e1af62be218178ca0df348a8b9ea6f not found: ID does not exist" containerID="e7aa8f07dd481c1734a9c89952ea1474e0e1af62be218178ca0df348a8b9ea6f" Mar 17 09:45:28 crc kubenswrapper[4813]: I0317 09:45:28.341352 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7aa8f07dd481c1734a9c89952ea1474e0e1af62be218178ca0df348a8b9ea6f"} err="failed to get container status \"e7aa8f07dd481c1734a9c89952ea1474e0e1af62be218178ca0df348a8b9ea6f\": rpc error: code = NotFound desc = could not find container \"e7aa8f07dd481c1734a9c89952ea1474e0e1af62be218178ca0df348a8b9ea6f\": container with ID starting with e7aa8f07dd481c1734a9c89952ea1474e0e1af62be218178ca0df348a8b9ea6f not found: ID does not exist" Mar 17 09:45:28 crc kubenswrapper[4813]: I0317 09:45:28.744571 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5541565-db4d-44af-9617-952d8077e9f1" path="/var/lib/kubelet/pods/f5541565-db4d-44af-9617-952d8077e9f1/volumes" Mar 17 09:45:47 crc kubenswrapper[4813]: I0317 09:45:47.411979 4813 generic.go:334] "Generic (PLEG): container finished" podID="5e4d8ae8-8ff6-4969-a349-813dde113094" containerID="2ae66e1640b8b81d8cebb9f7f86db75e3594f81fac08efa97c20dfc87d032ae9" exitCode=0 Mar 17 09:45:47 crc kubenswrapper[4813]: I0317 09:45:47.412068 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" event={"ID":"5e4d8ae8-8ff6-4969-a349-813dde113094","Type":"ContainerDied","Data":"2ae66e1640b8b81d8cebb9f7f86db75e3594f81fac08efa97c20dfc87d032ae9"} Mar 17 09:45:48 crc kubenswrapper[4813]: I0317 09:45:48.735685 4813 scope.go:117] "RemoveContainer" containerID="6abc6fdbecc5d29e40bf98aa15092c6a91d5d8667c098322c7987cf7b1896af4" Mar 17 09:45:48 crc kubenswrapper[4813]: I0317 09:45:48.891644 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.033014 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mn8bl\" (UniqueName: \"kubernetes.io/projected/5e4d8ae8-8ff6-4969-a349-813dde113094-kube-api-access-mn8bl\") pod \"5e4d8ae8-8ff6-4969-a349-813dde113094\" (UID: \"5e4d8ae8-8ff6-4969-a349-813dde113094\") " Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.033212 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5e4d8ae8-8ff6-4969-a349-813dde113094-ovncontroller-config-0\") pod \"5e4d8ae8-8ff6-4969-a349-813dde113094\" (UID: \"5e4d8ae8-8ff6-4969-a349-813dde113094\") " Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.033309 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e4d8ae8-8ff6-4969-a349-813dde113094-ssh-key-openstack-edpm-ipam\") pod \"5e4d8ae8-8ff6-4969-a349-813dde113094\" (UID: \"5e4d8ae8-8ff6-4969-a349-813dde113094\") " Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.033376 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e4d8ae8-8ff6-4969-a349-813dde113094-ovn-combined-ca-bundle\") pod \"5e4d8ae8-8ff6-4969-a349-813dde113094\" (UID: \"5e4d8ae8-8ff6-4969-a349-813dde113094\") " Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.033539 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e4d8ae8-8ff6-4969-a349-813dde113094-inventory\") pod \"5e4d8ae8-8ff6-4969-a349-813dde113094\" (UID: \"5e4d8ae8-8ff6-4969-a349-813dde113094\") " Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.047883 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e4d8ae8-8ff6-4969-a349-813dde113094-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "5e4d8ae8-8ff6-4969-a349-813dde113094" (UID: "5e4d8ae8-8ff6-4969-a349-813dde113094"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.047910 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e4d8ae8-8ff6-4969-a349-813dde113094-kube-api-access-mn8bl" (OuterVolumeSpecName: "kube-api-access-mn8bl") pod "5e4d8ae8-8ff6-4969-a349-813dde113094" (UID: "5e4d8ae8-8ff6-4969-a349-813dde113094"). InnerVolumeSpecName "kube-api-access-mn8bl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.057122 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e4d8ae8-8ff6-4969-a349-813dde113094-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "5e4d8ae8-8ff6-4969-a349-813dde113094" (UID: "5e4d8ae8-8ff6-4969-a349-813dde113094"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.061027 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e4d8ae8-8ff6-4969-a349-813dde113094-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "5e4d8ae8-8ff6-4969-a349-813dde113094" (UID: "5e4d8ae8-8ff6-4969-a349-813dde113094"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.070298 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e4d8ae8-8ff6-4969-a349-813dde113094-inventory" (OuterVolumeSpecName: "inventory") pod "5e4d8ae8-8ff6-4969-a349-813dde113094" (UID: "5e4d8ae8-8ff6-4969-a349-813dde113094"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.135775 4813 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5e4d8ae8-8ff6-4969-a349-813dde113094-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.136060 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/5e4d8ae8-8ff6-4969-a349-813dde113094-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.136154 4813 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e4d8ae8-8ff6-4969-a349-813dde113094-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.136231 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5e4d8ae8-8ff6-4969-a349-813dde113094-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.136314 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mn8bl\" (UniqueName: \"kubernetes.io/projected/5e4d8ae8-8ff6-4969-a349-813dde113094-kube-api-access-mn8bl\") on node \"crc\" DevicePath \"\"" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.434656 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" event={"ID":"5e4d8ae8-8ff6-4969-a349-813dde113094","Type":"ContainerDied","Data":"3562c103ce5e48a543271c2dd8b84759c5af9f9d252ca7c3ca51c5a1971aa447"} Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.434717 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3562c103ce5e48a543271c2dd8b84759c5af9f9d252ca7c3ca51c5a1971aa447" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.434879 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-c5ttr" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.536426 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj"] Mar 17 09:45:49 crc kubenswrapper[4813]: E0317 09:45:49.536797 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cadc4299-3256-40ba-b78c-8f863aff4254" containerName="extract-utilities" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.536808 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="cadc4299-3256-40ba-b78c-8f863aff4254" containerName="extract-utilities" Mar 17 09:45:49 crc kubenswrapper[4813]: E0317 09:45:49.536826 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e4d8ae8-8ff6-4969-a349-813dde113094" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.536832 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e4d8ae8-8ff6-4969-a349-813dde113094" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 17 09:45:49 crc kubenswrapper[4813]: E0317 09:45:49.536843 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5541565-db4d-44af-9617-952d8077e9f1" containerName="registry-server" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.536849 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5541565-db4d-44af-9617-952d8077e9f1" containerName="registry-server" Mar 17 09:45:49 crc kubenswrapper[4813]: E0317 09:45:49.536860 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cadc4299-3256-40ba-b78c-8f863aff4254" containerName="registry-server" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.536866 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="cadc4299-3256-40ba-b78c-8f863aff4254" containerName="registry-server" Mar 17 09:45:49 crc kubenswrapper[4813]: E0317 09:45:49.536882 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5541565-db4d-44af-9617-952d8077e9f1" containerName="extract-content" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.536888 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5541565-db4d-44af-9617-952d8077e9f1" containerName="extract-content" Mar 17 09:45:49 crc kubenswrapper[4813]: E0317 09:45:49.536903 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cadc4299-3256-40ba-b78c-8f863aff4254" containerName="extract-content" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.536909 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="cadc4299-3256-40ba-b78c-8f863aff4254" containerName="extract-content" Mar 17 09:45:49 crc kubenswrapper[4813]: E0317 09:45:49.536921 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5541565-db4d-44af-9617-952d8077e9f1" containerName="extract-utilities" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.536926 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5541565-db4d-44af-9617-952d8077e9f1" containerName="extract-utilities" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.537085 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5541565-db4d-44af-9617-952d8077e9f1" containerName="registry-server" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.537100 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="cadc4299-3256-40ba-b78c-8f863aff4254" containerName="registry-server" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.537113 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e4d8ae8-8ff6-4969-a349-813dde113094" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.537699 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.540955 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.541025 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.541115 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.541220 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.541398 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hj9h4" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.541433 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.569051 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj"] Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.643737 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.643828 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.644037 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.644114 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.644260 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-clxf6\" (UniqueName: \"kubernetes.io/projected/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-kube-api-access-clxf6\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.644389 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:45:49 crc kubenswrapper[4813]: E0317 09:45:49.657582 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e4d8ae8_8ff6_4969_a349_813dde113094.slice/crio-3562c103ce5e48a543271c2dd8b84759c5af9f9d252ca7c3ca51c5a1971aa447\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e4d8ae8_8ff6_4969_a349_813dde113094.slice\": RecentStats: unable to find data in memory cache]" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.747715 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.747779 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-clxf6\" (UniqueName: \"kubernetes.io/projected/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-kube-api-access-clxf6\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.747817 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.747917 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.747951 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.747992 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.756307 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.758097 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.758821 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.759402 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.769144 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-ssh-key-openstack-edpm-ipam\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.783330 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-clxf6\" (UniqueName: \"kubernetes.io/projected/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-kube-api-access-clxf6\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:45:49 crc kubenswrapper[4813]: I0317 09:45:49.858934 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:45:50 crc kubenswrapper[4813]: I0317 09:45:50.375350 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj"] Mar 17 09:45:50 crc kubenswrapper[4813]: I0317 09:45:50.445315 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" event={"ID":"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab","Type":"ContainerStarted","Data":"36c38f4e7adab6da8f78257f2cff57b11c99b0ce6c6d0fe3974e639ac3c167cd"} Mar 17 09:45:51 crc kubenswrapper[4813]: I0317 09:45:51.455260 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" event={"ID":"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab","Type":"ContainerStarted","Data":"99501eba6315d03311f0ae52c507ef2fede24d64e6b1ba0073bc44b8abfe2a0f"} Mar 17 09:45:51 crc kubenswrapper[4813]: I0317 09:45:51.471862 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" podStartSLOduration=1.9828404659999999 podStartE2EDuration="2.471844534s" podCreationTimestamp="2026-03-17 09:45:49 +0000 UTC" firstStartedPulling="2026-03-17 09:45:50.380238366 +0000 UTC m=+2172.481041875" lastFinishedPulling="2026-03-17 09:45:50.869242454 +0000 UTC m=+2172.970045943" observedRunningTime="2026-03-17 09:45:51.469170031 +0000 UTC m=+2173.569973530" watchObservedRunningTime="2026-03-17 09:45:51.471844534 +0000 UTC m=+2173.572648033" Mar 17 09:45:55 crc kubenswrapper[4813]: I0317 09:45:55.553695 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-q2m6c"] Mar 17 09:45:55 crc kubenswrapper[4813]: I0317 09:45:55.555823 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q2m6c" Mar 17 09:45:55 crc kubenswrapper[4813]: I0317 09:45:55.576308 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q2m6c"] Mar 17 09:45:55 crc kubenswrapper[4813]: I0317 09:45:55.667725 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqkjz\" (UniqueName: \"kubernetes.io/projected/8840c430-4827-4da1-baec-38cff419ed36-kube-api-access-sqkjz\") pod \"certified-operators-q2m6c\" (UID: \"8840c430-4827-4da1-baec-38cff419ed36\") " pod="openshift-marketplace/certified-operators-q2m6c" Mar 17 09:45:55 crc kubenswrapper[4813]: I0317 09:45:55.667830 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8840c430-4827-4da1-baec-38cff419ed36-catalog-content\") pod \"certified-operators-q2m6c\" (UID: \"8840c430-4827-4da1-baec-38cff419ed36\") " pod="openshift-marketplace/certified-operators-q2m6c" Mar 17 09:45:55 crc kubenswrapper[4813]: I0317 09:45:55.668002 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8840c430-4827-4da1-baec-38cff419ed36-utilities\") pod \"certified-operators-q2m6c\" (UID: \"8840c430-4827-4da1-baec-38cff419ed36\") " pod="openshift-marketplace/certified-operators-q2m6c" Mar 17 09:45:55 crc kubenswrapper[4813]: I0317 09:45:55.770182 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8840c430-4827-4da1-baec-38cff419ed36-utilities\") pod \"certified-operators-q2m6c\" (UID: \"8840c430-4827-4da1-baec-38cff419ed36\") " pod="openshift-marketplace/certified-operators-q2m6c" Mar 17 09:45:55 crc kubenswrapper[4813]: I0317 09:45:55.770516 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqkjz\" (UniqueName: \"kubernetes.io/projected/8840c430-4827-4da1-baec-38cff419ed36-kube-api-access-sqkjz\") pod \"certified-operators-q2m6c\" (UID: \"8840c430-4827-4da1-baec-38cff419ed36\") " pod="openshift-marketplace/certified-operators-q2m6c" Mar 17 09:45:55 crc kubenswrapper[4813]: I0317 09:45:55.770825 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8840c430-4827-4da1-baec-38cff419ed36-utilities\") pod \"certified-operators-q2m6c\" (UID: \"8840c430-4827-4da1-baec-38cff419ed36\") " pod="openshift-marketplace/certified-operators-q2m6c" Mar 17 09:45:55 crc kubenswrapper[4813]: I0317 09:45:55.771145 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8840c430-4827-4da1-baec-38cff419ed36-catalog-content\") pod \"certified-operators-q2m6c\" (UID: \"8840c430-4827-4da1-baec-38cff419ed36\") " pod="openshift-marketplace/certified-operators-q2m6c" Mar 17 09:45:55 crc kubenswrapper[4813]: I0317 09:45:55.771695 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8840c430-4827-4da1-baec-38cff419ed36-catalog-content\") pod \"certified-operators-q2m6c\" (UID: \"8840c430-4827-4da1-baec-38cff419ed36\") " pod="openshift-marketplace/certified-operators-q2m6c" Mar 17 09:45:55 crc kubenswrapper[4813]: I0317 09:45:55.800981 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqkjz\" (UniqueName: \"kubernetes.io/projected/8840c430-4827-4da1-baec-38cff419ed36-kube-api-access-sqkjz\") pod \"certified-operators-q2m6c\" (UID: \"8840c430-4827-4da1-baec-38cff419ed36\") " pod="openshift-marketplace/certified-operators-q2m6c" Mar 17 09:45:55 crc kubenswrapper[4813]: I0317 09:45:55.888266 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-q2m6c" Mar 17 09:45:56 crc kubenswrapper[4813]: I0317 09:45:56.523694 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q2m6c"] Mar 17 09:45:57 crc kubenswrapper[4813]: I0317 09:45:57.520133 4813 generic.go:334] "Generic (PLEG): container finished" podID="8840c430-4827-4da1-baec-38cff419ed36" containerID="d7abd52ec2d560022998bb267bbdf2f86e8112d0a0ae9b697a5d871af913e0d2" exitCode=0 Mar 17 09:45:57 crc kubenswrapper[4813]: I0317 09:45:57.520267 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2m6c" event={"ID":"8840c430-4827-4da1-baec-38cff419ed36","Type":"ContainerDied","Data":"d7abd52ec2d560022998bb267bbdf2f86e8112d0a0ae9b697a5d871af913e0d2"} Mar 17 09:45:57 crc kubenswrapper[4813]: I0317 09:45:57.520525 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2m6c" event={"ID":"8840c430-4827-4da1-baec-38cff419ed36","Type":"ContainerStarted","Data":"37627eb67358b3d5a5f5c1e207c43b80ca491a29e5325073d6fe28a738ce188e"} Mar 17 09:46:00 crc kubenswrapper[4813]: I0317 09:46:00.142915 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562346-rrzcl"] Mar 17 09:46:00 crc kubenswrapper[4813]: I0317 09:46:00.144728 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562346-rrzcl" Mar 17 09:46:00 crc kubenswrapper[4813]: I0317 09:46:00.147106 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:46:00 crc kubenswrapper[4813]: I0317 09:46:00.148496 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:46:00 crc kubenswrapper[4813]: I0317 09:46:00.151794 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562346-rrzcl"] Mar 17 09:46:00 crc kubenswrapper[4813]: I0317 09:46:00.154528 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:46:00 crc kubenswrapper[4813]: I0317 09:46:00.174812 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pdkpl\" (UniqueName: \"kubernetes.io/projected/4545fea0-ba23-4b1a-9640-bec6c63070dd-kube-api-access-pdkpl\") pod \"auto-csr-approver-29562346-rrzcl\" (UID: \"4545fea0-ba23-4b1a-9640-bec6c63070dd\") " pod="openshift-infra/auto-csr-approver-29562346-rrzcl" Mar 17 09:46:00 crc kubenswrapper[4813]: I0317 09:46:00.276534 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pdkpl\" (UniqueName: \"kubernetes.io/projected/4545fea0-ba23-4b1a-9640-bec6c63070dd-kube-api-access-pdkpl\") pod \"auto-csr-approver-29562346-rrzcl\" (UID: \"4545fea0-ba23-4b1a-9640-bec6c63070dd\") " pod="openshift-infra/auto-csr-approver-29562346-rrzcl" Mar 17 09:46:00 crc kubenswrapper[4813]: I0317 09:46:00.298858 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pdkpl\" (UniqueName: \"kubernetes.io/projected/4545fea0-ba23-4b1a-9640-bec6c63070dd-kube-api-access-pdkpl\") pod \"auto-csr-approver-29562346-rrzcl\" (UID: \"4545fea0-ba23-4b1a-9640-bec6c63070dd\") " pod="openshift-infra/auto-csr-approver-29562346-rrzcl" Mar 17 09:46:00 crc kubenswrapper[4813]: I0317 09:46:00.464371 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562346-rrzcl" Mar 17 09:46:02 crc kubenswrapper[4813]: I0317 09:46:02.420556 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562346-rrzcl"] Mar 17 09:46:02 crc kubenswrapper[4813]: I0317 09:46:02.575715 4813 generic.go:334] "Generic (PLEG): container finished" podID="8840c430-4827-4da1-baec-38cff419ed36" containerID="55583d00ae58c045ba388d0f4c362c42e8e3f18f0fd96279563d587b68a008f9" exitCode=0 Mar 17 09:46:02 crc kubenswrapper[4813]: I0317 09:46:02.575828 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2m6c" event={"ID":"8840c430-4827-4da1-baec-38cff419ed36","Type":"ContainerDied","Data":"55583d00ae58c045ba388d0f4c362c42e8e3f18f0fd96279563d587b68a008f9"} Mar 17 09:46:02 crc kubenswrapper[4813]: I0317 09:46:02.577829 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562346-rrzcl" event={"ID":"4545fea0-ba23-4b1a-9640-bec6c63070dd","Type":"ContainerStarted","Data":"628567c57baa641fa7b97fa36e143e397a56fe3843a5ca3d4b84eb77b4d625be"} Mar 17 09:46:03 crc kubenswrapper[4813]: I0317 09:46:03.604438 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-q2m6c" event={"ID":"8840c430-4827-4da1-baec-38cff419ed36","Type":"ContainerStarted","Data":"7dd45b299afc2f10863d20157e43adc5830877bd8143ef186952cabc59e96200"} Mar 17 09:46:03 crc kubenswrapper[4813]: I0317 09:46:03.620569 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-q2m6c" podStartSLOduration=3.147059301 podStartE2EDuration="8.620554467s" podCreationTimestamp="2026-03-17 09:45:55 +0000 UTC" firstStartedPulling="2026-03-17 09:45:57.522265029 +0000 UTC m=+2179.623068538" lastFinishedPulling="2026-03-17 09:46:02.995760195 +0000 UTC m=+2185.096563704" observedRunningTime="2026-03-17 09:46:03.618007788 +0000 UTC m=+2185.718811287" watchObservedRunningTime="2026-03-17 09:46:03.620554467 +0000 UTC m=+2185.721357966" Mar 17 09:46:04 crc kubenswrapper[4813]: I0317 09:46:04.616277 4813 generic.go:334] "Generic (PLEG): container finished" podID="4545fea0-ba23-4b1a-9640-bec6c63070dd" containerID="5f6299aaaf5f21de7750f72626a9708027e7ede92ce6871882347c5aa616330f" exitCode=0 Mar 17 09:46:04 crc kubenswrapper[4813]: I0317 09:46:04.616353 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562346-rrzcl" event={"ID":"4545fea0-ba23-4b1a-9640-bec6c63070dd","Type":"ContainerDied","Data":"5f6299aaaf5f21de7750f72626a9708027e7ede92ce6871882347c5aa616330f"} Mar 17 09:46:05 crc kubenswrapper[4813]: I0317 09:46:05.892681 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-q2m6c" Mar 17 09:46:05 crc kubenswrapper[4813]: I0317 09:46:05.894017 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-q2m6c" Mar 17 09:46:05 crc kubenswrapper[4813]: I0317 09:46:05.962400 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-q2m6c" Mar 17 09:46:06 crc kubenswrapper[4813]: I0317 09:46:06.057528 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562346-rrzcl" Mar 17 09:46:06 crc kubenswrapper[4813]: I0317 09:46:06.204443 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pdkpl\" (UniqueName: \"kubernetes.io/projected/4545fea0-ba23-4b1a-9640-bec6c63070dd-kube-api-access-pdkpl\") pod \"4545fea0-ba23-4b1a-9640-bec6c63070dd\" (UID: \"4545fea0-ba23-4b1a-9640-bec6c63070dd\") " Mar 17 09:46:06 crc kubenswrapper[4813]: I0317 09:46:06.212699 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4545fea0-ba23-4b1a-9640-bec6c63070dd-kube-api-access-pdkpl" (OuterVolumeSpecName: "kube-api-access-pdkpl") pod "4545fea0-ba23-4b1a-9640-bec6c63070dd" (UID: "4545fea0-ba23-4b1a-9640-bec6c63070dd"). InnerVolumeSpecName "kube-api-access-pdkpl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:46:06 crc kubenswrapper[4813]: I0317 09:46:06.308804 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pdkpl\" (UniqueName: \"kubernetes.io/projected/4545fea0-ba23-4b1a-9640-bec6c63070dd-kube-api-access-pdkpl\") on node \"crc\" DevicePath \"\"" Mar 17 09:46:06 crc kubenswrapper[4813]: I0317 09:46:06.642734 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562346-rrzcl" event={"ID":"4545fea0-ba23-4b1a-9640-bec6c63070dd","Type":"ContainerDied","Data":"628567c57baa641fa7b97fa36e143e397a56fe3843a5ca3d4b84eb77b4d625be"} Mar 17 09:46:06 crc kubenswrapper[4813]: I0317 09:46:06.642793 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="628567c57baa641fa7b97fa36e143e397a56fe3843a5ca3d4b84eb77b4d625be" Mar 17 09:46:06 crc kubenswrapper[4813]: I0317 09:46:06.642759 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562346-rrzcl" Mar 17 09:46:07 crc kubenswrapper[4813]: I0317 09:46:07.130676 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562340-cw8rp"] Mar 17 09:46:07 crc kubenswrapper[4813]: I0317 09:46:07.140294 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562340-cw8rp"] Mar 17 09:46:08 crc kubenswrapper[4813]: I0317 09:46:08.752041 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b5d72c0-ddd6-4c41-8845-7052c340a046" path="/var/lib/kubelet/pods/8b5d72c0-ddd6-4c41-8845-7052c340a046/volumes" Mar 17 09:46:15 crc kubenswrapper[4813]: I0317 09:46:15.973933 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-q2m6c" Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.135458 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-q2m6c"] Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.159863 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j7m85"] Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.160303 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-j7m85" podUID="a15c3d28-d8aa-4d38-a180-420c05345960" containerName="registry-server" containerID="cri-o://38de3e851b2a82e6206e5adf6ce5348414401b6d3e6657ac5ef2249d45f2a213" gracePeriod=2 Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.623223 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j7m85" Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.647445 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a15c3d28-d8aa-4d38-a180-420c05345960-catalog-content\") pod \"a15c3d28-d8aa-4d38-a180-420c05345960\" (UID: \"a15c3d28-d8aa-4d38-a180-420c05345960\") " Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.647512 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h7hkx\" (UniqueName: \"kubernetes.io/projected/a15c3d28-d8aa-4d38-a180-420c05345960-kube-api-access-h7hkx\") pod \"a15c3d28-d8aa-4d38-a180-420c05345960\" (UID: \"a15c3d28-d8aa-4d38-a180-420c05345960\") " Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.647586 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a15c3d28-d8aa-4d38-a180-420c05345960-utilities\") pod \"a15c3d28-d8aa-4d38-a180-420c05345960\" (UID: \"a15c3d28-d8aa-4d38-a180-420c05345960\") " Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.648179 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a15c3d28-d8aa-4d38-a180-420c05345960-utilities" (OuterVolumeSpecName: "utilities") pod "a15c3d28-d8aa-4d38-a180-420c05345960" (UID: "a15c3d28-d8aa-4d38-a180-420c05345960"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.653569 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a15c3d28-d8aa-4d38-a180-420c05345960-kube-api-access-h7hkx" (OuterVolumeSpecName: "kube-api-access-h7hkx") pod "a15c3d28-d8aa-4d38-a180-420c05345960" (UID: "a15c3d28-d8aa-4d38-a180-420c05345960"). InnerVolumeSpecName "kube-api-access-h7hkx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.705915 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a15c3d28-d8aa-4d38-a180-420c05345960-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a15c3d28-d8aa-4d38-a180-420c05345960" (UID: "a15c3d28-d8aa-4d38-a180-420c05345960"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.749897 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a15c3d28-d8aa-4d38-a180-420c05345960-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.749918 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h7hkx\" (UniqueName: \"kubernetes.io/projected/a15c3d28-d8aa-4d38-a180-420c05345960-kube-api-access-h7hkx\") on node \"crc\" DevicePath \"\"" Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.749929 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a15c3d28-d8aa-4d38-a180-420c05345960-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.778932 4813 generic.go:334] "Generic (PLEG): container finished" podID="a15c3d28-d8aa-4d38-a180-420c05345960" containerID="38de3e851b2a82e6206e5adf6ce5348414401b6d3e6657ac5ef2249d45f2a213" exitCode=0 Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.778996 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-j7m85" Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.779025 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7m85" event={"ID":"a15c3d28-d8aa-4d38-a180-420c05345960","Type":"ContainerDied","Data":"38de3e851b2a82e6206e5adf6ce5348414401b6d3e6657ac5ef2249d45f2a213"} Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.779062 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-j7m85" event={"ID":"a15c3d28-d8aa-4d38-a180-420c05345960","Type":"ContainerDied","Data":"235123a2e4e9b20c943afb86d8c6d8956d0d581a86d85bbcfe7615b3b049e46a"} Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.779081 4813 scope.go:117] "RemoveContainer" containerID="38de3e851b2a82e6206e5adf6ce5348414401b6d3e6657ac5ef2249d45f2a213" Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.800085 4813 scope.go:117] "RemoveContainer" containerID="0d4083d119d8c44805c93913a7f730a5c5bd90393faa5d3a37a1476d8d237c01" Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.804706 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-j7m85"] Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.819344 4813 scope.go:117] "RemoveContainer" containerID="5e5caf98ec1b4ebbec66e61c1ee07e96b7274f80315ba8aed847348d5c9c8109" Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.821805 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-j7m85"] Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.874510 4813 scope.go:117] "RemoveContainer" containerID="38de3e851b2a82e6206e5adf6ce5348414401b6d3e6657ac5ef2249d45f2a213" Mar 17 09:46:16 crc kubenswrapper[4813]: E0317 09:46:16.875064 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"38de3e851b2a82e6206e5adf6ce5348414401b6d3e6657ac5ef2249d45f2a213\": container with ID starting with 38de3e851b2a82e6206e5adf6ce5348414401b6d3e6657ac5ef2249d45f2a213 not found: ID does not exist" containerID="38de3e851b2a82e6206e5adf6ce5348414401b6d3e6657ac5ef2249d45f2a213" Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.875098 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"38de3e851b2a82e6206e5adf6ce5348414401b6d3e6657ac5ef2249d45f2a213"} err="failed to get container status \"38de3e851b2a82e6206e5adf6ce5348414401b6d3e6657ac5ef2249d45f2a213\": rpc error: code = NotFound desc = could not find container \"38de3e851b2a82e6206e5adf6ce5348414401b6d3e6657ac5ef2249d45f2a213\": container with ID starting with 38de3e851b2a82e6206e5adf6ce5348414401b6d3e6657ac5ef2249d45f2a213 not found: ID does not exist" Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.875127 4813 scope.go:117] "RemoveContainer" containerID="0d4083d119d8c44805c93913a7f730a5c5bd90393faa5d3a37a1476d8d237c01" Mar 17 09:46:16 crc kubenswrapper[4813]: E0317 09:46:16.875483 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0d4083d119d8c44805c93913a7f730a5c5bd90393faa5d3a37a1476d8d237c01\": container with ID starting with 0d4083d119d8c44805c93913a7f730a5c5bd90393faa5d3a37a1476d8d237c01 not found: ID does not exist" containerID="0d4083d119d8c44805c93913a7f730a5c5bd90393faa5d3a37a1476d8d237c01" Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.875509 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0d4083d119d8c44805c93913a7f730a5c5bd90393faa5d3a37a1476d8d237c01"} err="failed to get container status \"0d4083d119d8c44805c93913a7f730a5c5bd90393faa5d3a37a1476d8d237c01\": rpc error: code = NotFound desc = could not find container \"0d4083d119d8c44805c93913a7f730a5c5bd90393faa5d3a37a1476d8d237c01\": container with ID starting with 0d4083d119d8c44805c93913a7f730a5c5bd90393faa5d3a37a1476d8d237c01 not found: ID does not exist" Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.875528 4813 scope.go:117] "RemoveContainer" containerID="5e5caf98ec1b4ebbec66e61c1ee07e96b7274f80315ba8aed847348d5c9c8109" Mar 17 09:46:16 crc kubenswrapper[4813]: E0317 09:46:16.875842 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e5caf98ec1b4ebbec66e61c1ee07e96b7274f80315ba8aed847348d5c9c8109\": container with ID starting with 5e5caf98ec1b4ebbec66e61c1ee07e96b7274f80315ba8aed847348d5c9c8109 not found: ID does not exist" containerID="5e5caf98ec1b4ebbec66e61c1ee07e96b7274f80315ba8aed847348d5c9c8109" Mar 17 09:46:16 crc kubenswrapper[4813]: I0317 09:46:16.875886 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e5caf98ec1b4ebbec66e61c1ee07e96b7274f80315ba8aed847348d5c9c8109"} err="failed to get container status \"5e5caf98ec1b4ebbec66e61c1ee07e96b7274f80315ba8aed847348d5c9c8109\": rpc error: code = NotFound desc = could not find container \"5e5caf98ec1b4ebbec66e61c1ee07e96b7274f80315ba8aed847348d5c9c8109\": container with ID starting with 5e5caf98ec1b4ebbec66e61c1ee07e96b7274f80315ba8aed847348d5c9c8109 not found: ID does not exist" Mar 17 09:46:18 crc kubenswrapper[4813]: I0317 09:46:18.742915 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a15c3d28-d8aa-4d38-a180-420c05345960" path="/var/lib/kubelet/pods/a15c3d28-d8aa-4d38-a180-420c05345960/volumes" Mar 17 09:46:35 crc kubenswrapper[4813]: I0317 09:46:35.789370 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openstack/swift-proxy-8557f6579f-shvsw" podUID="e79b20b0-682e-4d16-bb56-64f0c4ec0202" containerName="proxy-server" probeResult="failure" output="HTTP probe failed with statuscode: 502" Mar 17 09:46:42 crc kubenswrapper[4813]: I0317 09:46:42.044896 4813 generic.go:334] "Generic (PLEG): container finished" podID="0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab" containerID="99501eba6315d03311f0ae52c507ef2fede24d64e6b1ba0073bc44b8abfe2a0f" exitCode=0 Mar 17 09:46:42 crc kubenswrapper[4813]: I0317 09:46:42.044963 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" event={"ID":"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab","Type":"ContainerDied","Data":"99501eba6315d03311f0ae52c507ef2fede24d64e6b1ba0073bc44b8abfe2a0f"} Mar 17 09:46:43 crc kubenswrapper[4813]: I0317 09:46:43.608244 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:46:43 crc kubenswrapper[4813]: I0317 09:46:43.687792 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-nova-metadata-neutron-config-0\") pod \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " Mar 17 09:46:43 crc kubenswrapper[4813]: I0317 09:46:43.687917 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-ssh-key-openstack-edpm-ipam\") pod \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " Mar 17 09:46:43 crc kubenswrapper[4813]: I0317 09:46:43.687965 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-neutron-metadata-combined-ca-bundle\") pod \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " Mar 17 09:46:43 crc kubenswrapper[4813]: I0317 09:46:43.688065 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-clxf6\" (UniqueName: \"kubernetes.io/projected/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-kube-api-access-clxf6\") pod \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " Mar 17 09:46:43 crc kubenswrapper[4813]: I0317 09:46:43.688140 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-inventory\") pod \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " Mar 17 09:46:43 crc kubenswrapper[4813]: I0317 09:46:43.688223 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-neutron-ovn-metadata-agent-neutron-config-0\") pod \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\" (UID: \"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab\") " Mar 17 09:46:43 crc kubenswrapper[4813]: I0317 09:46:43.693914 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-kube-api-access-clxf6" (OuterVolumeSpecName: "kube-api-access-clxf6") pod "0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab" (UID: "0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab"). InnerVolumeSpecName "kube-api-access-clxf6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:46:43 crc kubenswrapper[4813]: I0317 09:46:43.697778 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab" (UID: "0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:46:43 crc kubenswrapper[4813]: I0317 09:46:43.723231 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-inventory" (OuterVolumeSpecName: "inventory") pod "0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab" (UID: "0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:46:43 crc kubenswrapper[4813]: I0317 09:46:43.723793 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab" (UID: "0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:46:43 crc kubenswrapper[4813]: I0317 09:46:43.733082 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab" (UID: "0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:46:43 crc kubenswrapper[4813]: I0317 09:46:43.744277 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab" (UID: "0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:46:43 crc kubenswrapper[4813]: I0317 09:46:43.790922 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 09:46:43 crc kubenswrapper[4813]: I0317 09:46:43.790961 4813 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:46:43 crc kubenswrapper[4813]: I0317 09:46:43.790976 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-clxf6\" (UniqueName: \"kubernetes.io/projected/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-kube-api-access-clxf6\") on node \"crc\" DevicePath \"\"" Mar 17 09:46:43 crc kubenswrapper[4813]: I0317 09:46:43.790990 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 09:46:43 crc kubenswrapper[4813]: I0317 09:46:43.791003 4813 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:46:43 crc kubenswrapper[4813]: I0317 09:46:43.791018 4813 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.070252 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" event={"ID":"0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab","Type":"ContainerDied","Data":"36c38f4e7adab6da8f78257f2cff57b11c99b0ce6c6d0fe3974e639ac3c167cd"} Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.070313 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36c38f4e7adab6da8f78257f2cff57b11c99b0ce6c6d0fe3974e639ac3c167cd" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.070360 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.235591 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8"] Mar 17 09:46:44 crc kubenswrapper[4813]: E0317 09:46:44.236063 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a15c3d28-d8aa-4d38-a180-420c05345960" containerName="registry-server" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.236090 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a15c3d28-d8aa-4d38-a180-420c05345960" containerName="registry-server" Mar 17 09:46:44 crc kubenswrapper[4813]: E0317 09:46:44.236108 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a15c3d28-d8aa-4d38-a180-420c05345960" containerName="extract-utilities" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.236117 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a15c3d28-d8aa-4d38-a180-420c05345960" containerName="extract-utilities" Mar 17 09:46:44 crc kubenswrapper[4813]: E0317 09:46:44.236141 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4545fea0-ba23-4b1a-9640-bec6c63070dd" containerName="oc" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.236149 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4545fea0-ba23-4b1a-9640-bec6c63070dd" containerName="oc" Mar 17 09:46:44 crc kubenswrapper[4813]: E0317 09:46:44.236173 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a15c3d28-d8aa-4d38-a180-420c05345960" containerName="extract-content" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.236182 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a15c3d28-d8aa-4d38-a180-420c05345960" containerName="extract-content" Mar 17 09:46:44 crc kubenswrapper[4813]: E0317 09:46:44.236212 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.236224 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.236443 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="4545fea0-ba23-4b1a-9640-bec6c63070dd" containerName="oc" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.236462 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a15c3d28-d8aa-4d38-a180-420c05345960" containerName="registry-server" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.236487 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.238176 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.241581 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.241880 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.242362 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.242859 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hj9h4" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.245260 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.256052 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8"] Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.300904 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxhnb\" (UniqueName: \"kubernetes.io/projected/71e0074f-5ee2-4f59-9184-c23495021bfd-kube-api-access-bxhnb\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8\" (UID: \"71e0074f-5ee2-4f59-9184-c23495021bfd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.300984 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8\" (UID: \"71e0074f-5ee2-4f59-9184-c23495021bfd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.301113 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8\" (UID: \"71e0074f-5ee2-4f59-9184-c23495021bfd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.301141 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8\" (UID: \"71e0074f-5ee2-4f59-9184-c23495021bfd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.301168 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8\" (UID: \"71e0074f-5ee2-4f59-9184-c23495021bfd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.402743 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8\" (UID: \"71e0074f-5ee2-4f59-9184-c23495021bfd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.402787 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8\" (UID: \"71e0074f-5ee2-4f59-9184-c23495021bfd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.402811 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8\" (UID: \"71e0074f-5ee2-4f59-9184-c23495021bfd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.402880 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxhnb\" (UniqueName: \"kubernetes.io/projected/71e0074f-5ee2-4f59-9184-c23495021bfd-kube-api-access-bxhnb\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8\" (UID: \"71e0074f-5ee2-4f59-9184-c23495021bfd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.402913 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8\" (UID: \"71e0074f-5ee2-4f59-9184-c23495021bfd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.408314 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8\" (UID: \"71e0074f-5ee2-4f59-9184-c23495021bfd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.409524 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-ssh-key-openstack-edpm-ipam\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8\" (UID: \"71e0074f-5ee2-4f59-9184-c23495021bfd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.410529 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8\" (UID: \"71e0074f-5ee2-4f59-9184-c23495021bfd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.411545 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8\" (UID: \"71e0074f-5ee2-4f59-9184-c23495021bfd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.424830 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxhnb\" (UniqueName: \"kubernetes.io/projected/71e0074f-5ee2-4f59-9184-c23495021bfd-kube-api-access-bxhnb\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8\" (UID: \"71e0074f-5ee2-4f59-9184-c23495021bfd\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" Mar 17 09:46:44 crc kubenswrapper[4813]: I0317 09:46:44.558506 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" Mar 17 09:46:45 crc kubenswrapper[4813]: I0317 09:46:45.083205 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8"] Mar 17 09:46:45 crc kubenswrapper[4813]: I0317 09:46:45.096975 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 09:46:46 crc kubenswrapper[4813]: I0317 09:46:46.088171 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" event={"ID":"71e0074f-5ee2-4f59-9184-c23495021bfd","Type":"ContainerStarted","Data":"8f6c8123ad3b3e604a8fa61380731e0b2bc2e024140d7b7186c3dcae3b49177a"} Mar 17 09:46:46 crc kubenswrapper[4813]: I0317 09:46:46.088585 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" event={"ID":"71e0074f-5ee2-4f59-9184-c23495021bfd","Type":"ContainerStarted","Data":"befc1780e63f4560aec113dc77461d850713071f980743276dc5aa119972c97e"} Mar 17 09:46:46 crc kubenswrapper[4813]: I0317 09:46:46.118363 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" podStartSLOduration=1.587968127 podStartE2EDuration="2.118341232s" podCreationTimestamp="2026-03-17 09:46:44 +0000 UTC" firstStartedPulling="2026-03-17 09:46:45.096351029 +0000 UTC m=+2227.197154568" lastFinishedPulling="2026-03-17 09:46:45.626724174 +0000 UTC m=+2227.727527673" observedRunningTime="2026-03-17 09:46:46.11078586 +0000 UTC m=+2228.211589369" watchObservedRunningTime="2026-03-17 09:46:46.118341232 +0000 UTC m=+2228.219144751" Mar 17 09:46:48 crc kubenswrapper[4813]: I0317 09:46:48.844758 4813 scope.go:117] "RemoveContainer" containerID="09b1e1e803fe358a61aa5bfed604dbac24f44be69505e18ae4d8b75ed810f988" Mar 17 09:47:14 crc kubenswrapper[4813]: I0317 09:47:14.114289 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:47:14 crc kubenswrapper[4813]: I0317 09:47:14.115031 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:47:44 crc kubenswrapper[4813]: I0317 09:47:44.113886 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:47:44 crc kubenswrapper[4813]: I0317 09:47:44.114692 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:48:00 crc kubenswrapper[4813]: I0317 09:48:00.160516 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562348-lt7kt"] Mar 17 09:48:00 crc kubenswrapper[4813]: I0317 09:48:00.163311 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562348-lt7kt" Mar 17 09:48:00 crc kubenswrapper[4813]: I0317 09:48:00.167410 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:48:00 crc kubenswrapper[4813]: I0317 09:48:00.167658 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:48:00 crc kubenswrapper[4813]: I0317 09:48:00.168512 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:48:00 crc kubenswrapper[4813]: I0317 09:48:00.171117 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562348-lt7kt"] Mar 17 09:48:00 crc kubenswrapper[4813]: I0317 09:48:00.195948 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld2h5\" (UniqueName: \"kubernetes.io/projected/ed42eba9-9fb3-4505-acd2-c552b16139b0-kube-api-access-ld2h5\") pod \"auto-csr-approver-29562348-lt7kt\" (UID: \"ed42eba9-9fb3-4505-acd2-c552b16139b0\") " pod="openshift-infra/auto-csr-approver-29562348-lt7kt" Mar 17 09:48:00 crc kubenswrapper[4813]: I0317 09:48:00.297801 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld2h5\" (UniqueName: \"kubernetes.io/projected/ed42eba9-9fb3-4505-acd2-c552b16139b0-kube-api-access-ld2h5\") pod \"auto-csr-approver-29562348-lt7kt\" (UID: \"ed42eba9-9fb3-4505-acd2-c552b16139b0\") " pod="openshift-infra/auto-csr-approver-29562348-lt7kt" Mar 17 09:48:00 crc kubenswrapper[4813]: I0317 09:48:00.318195 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld2h5\" (UniqueName: \"kubernetes.io/projected/ed42eba9-9fb3-4505-acd2-c552b16139b0-kube-api-access-ld2h5\") pod \"auto-csr-approver-29562348-lt7kt\" (UID: \"ed42eba9-9fb3-4505-acd2-c552b16139b0\") " pod="openshift-infra/auto-csr-approver-29562348-lt7kt" Mar 17 09:48:00 crc kubenswrapper[4813]: I0317 09:48:00.493743 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562348-lt7kt" Mar 17 09:48:00 crc kubenswrapper[4813]: I0317 09:48:00.976281 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562348-lt7kt"] Mar 17 09:48:01 crc kubenswrapper[4813]: I0317 09:48:01.924889 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562348-lt7kt" event={"ID":"ed42eba9-9fb3-4505-acd2-c552b16139b0","Type":"ContainerStarted","Data":"35dc30ad379640e509f68fcd7a8aa018c9948eb25fc616b8784530052d4f1374"} Mar 17 09:48:02 crc kubenswrapper[4813]: I0317 09:48:02.937117 4813 generic.go:334] "Generic (PLEG): container finished" podID="ed42eba9-9fb3-4505-acd2-c552b16139b0" containerID="93bbbd082fb016f1f3994a86f419e32766d9241ae7608f82c19120e91e144157" exitCode=0 Mar 17 09:48:02 crc kubenswrapper[4813]: I0317 09:48:02.937165 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562348-lt7kt" event={"ID":"ed42eba9-9fb3-4505-acd2-c552b16139b0","Type":"ContainerDied","Data":"93bbbd082fb016f1f3994a86f419e32766d9241ae7608f82c19120e91e144157"} Mar 17 09:48:04 crc kubenswrapper[4813]: I0317 09:48:04.383058 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562348-lt7kt" Mar 17 09:48:04 crc kubenswrapper[4813]: I0317 09:48:04.484134 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld2h5\" (UniqueName: \"kubernetes.io/projected/ed42eba9-9fb3-4505-acd2-c552b16139b0-kube-api-access-ld2h5\") pod \"ed42eba9-9fb3-4505-acd2-c552b16139b0\" (UID: \"ed42eba9-9fb3-4505-acd2-c552b16139b0\") " Mar 17 09:48:04 crc kubenswrapper[4813]: I0317 09:48:04.490853 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed42eba9-9fb3-4505-acd2-c552b16139b0-kube-api-access-ld2h5" (OuterVolumeSpecName: "kube-api-access-ld2h5") pod "ed42eba9-9fb3-4505-acd2-c552b16139b0" (UID: "ed42eba9-9fb3-4505-acd2-c552b16139b0"). InnerVolumeSpecName "kube-api-access-ld2h5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:48:04 crc kubenswrapper[4813]: I0317 09:48:04.586886 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld2h5\" (UniqueName: \"kubernetes.io/projected/ed42eba9-9fb3-4505-acd2-c552b16139b0-kube-api-access-ld2h5\") on node \"crc\" DevicePath \"\"" Mar 17 09:48:04 crc kubenswrapper[4813]: I0317 09:48:04.956314 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562348-lt7kt" event={"ID":"ed42eba9-9fb3-4505-acd2-c552b16139b0","Type":"ContainerDied","Data":"35dc30ad379640e509f68fcd7a8aa018c9948eb25fc616b8784530052d4f1374"} Mar 17 09:48:04 crc kubenswrapper[4813]: I0317 09:48:04.956350 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="35dc30ad379640e509f68fcd7a8aa018c9948eb25fc616b8784530052d4f1374" Mar 17 09:48:04 crc kubenswrapper[4813]: I0317 09:48:04.956426 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562348-lt7kt" Mar 17 09:48:05 crc kubenswrapper[4813]: I0317 09:48:05.467206 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562342-bq4l7"] Mar 17 09:48:05 crc kubenswrapper[4813]: I0317 09:48:05.477307 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562342-bq4l7"] Mar 17 09:48:06 crc kubenswrapper[4813]: I0317 09:48:06.751956 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b2f2eb2-09cd-422d-80a1-630d2811c390" path="/var/lib/kubelet/pods/0b2f2eb2-09cd-422d-80a1-630d2811c390/volumes" Mar 17 09:48:14 crc kubenswrapper[4813]: I0317 09:48:14.113798 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:48:14 crc kubenswrapper[4813]: I0317 09:48:14.114304 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:48:14 crc kubenswrapper[4813]: I0317 09:48:14.114353 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:48:14 crc kubenswrapper[4813]: I0317 09:48:14.115062 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4"} pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 09:48:14 crc kubenswrapper[4813]: I0317 09:48:14.115115 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" containerID="cri-o://5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" gracePeriod=600 Mar 17 09:48:14 crc kubenswrapper[4813]: E0317 09:48:14.240259 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:48:15 crc kubenswrapper[4813]: I0317 09:48:15.103040 4813 generic.go:334] "Generic (PLEG): container finished" podID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" exitCode=0 Mar 17 09:48:15 crc kubenswrapper[4813]: I0317 09:48:15.103132 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerDied","Data":"5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4"} Mar 17 09:48:15 crc kubenswrapper[4813]: I0317 09:48:15.103456 4813 scope.go:117] "RemoveContainer" containerID="886e3a29e582cbf7efd8021b4dad6057b4824ee4ecc8a4ac9ae12df4687c6449" Mar 17 09:48:15 crc kubenswrapper[4813]: I0317 09:48:15.104839 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:48:15 crc kubenswrapper[4813]: E0317 09:48:15.105458 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:48:20 crc kubenswrapper[4813]: I0317 09:48:20.019153 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h6wdl"] Mar 17 09:48:20 crc kubenswrapper[4813]: E0317 09:48:20.020281 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed42eba9-9fb3-4505-acd2-c552b16139b0" containerName="oc" Mar 17 09:48:20 crc kubenswrapper[4813]: I0317 09:48:20.020304 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed42eba9-9fb3-4505-acd2-c552b16139b0" containerName="oc" Mar 17 09:48:20 crc kubenswrapper[4813]: I0317 09:48:20.020874 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed42eba9-9fb3-4505-acd2-c552b16139b0" containerName="oc" Mar 17 09:48:20 crc kubenswrapper[4813]: I0317 09:48:20.023279 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h6wdl" Mar 17 09:48:20 crc kubenswrapper[4813]: I0317 09:48:20.044183 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h6wdl"] Mar 17 09:48:20 crc kubenswrapper[4813]: I0317 09:48:20.100283 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/561032e6-7a20-40de-a141-88b07c2ee5ef-catalog-content\") pod \"redhat-marketplace-h6wdl\" (UID: \"561032e6-7a20-40de-a141-88b07c2ee5ef\") " pod="openshift-marketplace/redhat-marketplace-h6wdl" Mar 17 09:48:20 crc kubenswrapper[4813]: I0317 09:48:20.100499 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/561032e6-7a20-40de-a141-88b07c2ee5ef-utilities\") pod \"redhat-marketplace-h6wdl\" (UID: \"561032e6-7a20-40de-a141-88b07c2ee5ef\") " pod="openshift-marketplace/redhat-marketplace-h6wdl" Mar 17 09:48:20 crc kubenswrapper[4813]: I0317 09:48:20.100531 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k77dq\" (UniqueName: \"kubernetes.io/projected/561032e6-7a20-40de-a141-88b07c2ee5ef-kube-api-access-k77dq\") pod \"redhat-marketplace-h6wdl\" (UID: \"561032e6-7a20-40de-a141-88b07c2ee5ef\") " pod="openshift-marketplace/redhat-marketplace-h6wdl" Mar 17 09:48:20 crc kubenswrapper[4813]: I0317 09:48:20.202188 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/561032e6-7a20-40de-a141-88b07c2ee5ef-utilities\") pod \"redhat-marketplace-h6wdl\" (UID: \"561032e6-7a20-40de-a141-88b07c2ee5ef\") " pod="openshift-marketplace/redhat-marketplace-h6wdl" Mar 17 09:48:20 crc kubenswrapper[4813]: I0317 09:48:20.202229 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k77dq\" (UniqueName: \"kubernetes.io/projected/561032e6-7a20-40de-a141-88b07c2ee5ef-kube-api-access-k77dq\") pod \"redhat-marketplace-h6wdl\" (UID: \"561032e6-7a20-40de-a141-88b07c2ee5ef\") " pod="openshift-marketplace/redhat-marketplace-h6wdl" Mar 17 09:48:20 crc kubenswrapper[4813]: I0317 09:48:20.202295 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/561032e6-7a20-40de-a141-88b07c2ee5ef-catalog-content\") pod \"redhat-marketplace-h6wdl\" (UID: \"561032e6-7a20-40de-a141-88b07c2ee5ef\") " pod="openshift-marketplace/redhat-marketplace-h6wdl" Mar 17 09:48:20 crc kubenswrapper[4813]: I0317 09:48:20.202747 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/561032e6-7a20-40de-a141-88b07c2ee5ef-utilities\") pod \"redhat-marketplace-h6wdl\" (UID: \"561032e6-7a20-40de-a141-88b07c2ee5ef\") " pod="openshift-marketplace/redhat-marketplace-h6wdl" Mar 17 09:48:20 crc kubenswrapper[4813]: I0317 09:48:20.202797 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/561032e6-7a20-40de-a141-88b07c2ee5ef-catalog-content\") pod \"redhat-marketplace-h6wdl\" (UID: \"561032e6-7a20-40de-a141-88b07c2ee5ef\") " pod="openshift-marketplace/redhat-marketplace-h6wdl" Mar 17 09:48:20 crc kubenswrapper[4813]: I0317 09:48:20.224706 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k77dq\" (UniqueName: \"kubernetes.io/projected/561032e6-7a20-40de-a141-88b07c2ee5ef-kube-api-access-k77dq\") pod \"redhat-marketplace-h6wdl\" (UID: \"561032e6-7a20-40de-a141-88b07c2ee5ef\") " pod="openshift-marketplace/redhat-marketplace-h6wdl" Mar 17 09:48:20 crc kubenswrapper[4813]: I0317 09:48:20.392589 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h6wdl" Mar 17 09:48:20 crc kubenswrapper[4813]: I0317 09:48:20.830853 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h6wdl"] Mar 17 09:48:20 crc kubenswrapper[4813]: W0317 09:48:20.835116 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod561032e6_7a20_40de_a141_88b07c2ee5ef.slice/crio-4a263e02580cbc1bd52cd87523865bc6d9d39c3d54645c30df23466f9bf8b76d WatchSource:0}: Error finding container 4a263e02580cbc1bd52cd87523865bc6d9d39c3d54645c30df23466f9bf8b76d: Status 404 returned error can't find the container with id 4a263e02580cbc1bd52cd87523865bc6d9d39c3d54645c30df23466f9bf8b76d Mar 17 09:48:21 crc kubenswrapper[4813]: I0317 09:48:21.167187 4813 generic.go:334] "Generic (PLEG): container finished" podID="561032e6-7a20-40de-a141-88b07c2ee5ef" containerID="0447f6b3327a1f75bcc942832fcb55fa3331bd34e952778f74c9d7754d62b14e" exitCode=0 Mar 17 09:48:21 crc kubenswrapper[4813]: I0317 09:48:21.167297 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h6wdl" event={"ID":"561032e6-7a20-40de-a141-88b07c2ee5ef","Type":"ContainerDied","Data":"0447f6b3327a1f75bcc942832fcb55fa3331bd34e952778f74c9d7754d62b14e"} Mar 17 09:48:21 crc kubenswrapper[4813]: I0317 09:48:21.167548 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h6wdl" event={"ID":"561032e6-7a20-40de-a141-88b07c2ee5ef","Type":"ContainerStarted","Data":"4a263e02580cbc1bd52cd87523865bc6d9d39c3d54645c30df23466f9bf8b76d"} Mar 17 09:48:23 crc kubenswrapper[4813]: I0317 09:48:23.185437 4813 generic.go:334] "Generic (PLEG): container finished" podID="561032e6-7a20-40de-a141-88b07c2ee5ef" containerID="f6b111c68dd60f004d5b9306150c269e2dcd53446715f73252dacd88b33c1a0b" exitCode=0 Mar 17 09:48:23 crc kubenswrapper[4813]: I0317 09:48:23.185638 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h6wdl" event={"ID":"561032e6-7a20-40de-a141-88b07c2ee5ef","Type":"ContainerDied","Data":"f6b111c68dd60f004d5b9306150c269e2dcd53446715f73252dacd88b33c1a0b"} Mar 17 09:48:24 crc kubenswrapper[4813]: I0317 09:48:24.200975 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h6wdl" event={"ID":"561032e6-7a20-40de-a141-88b07c2ee5ef","Type":"ContainerStarted","Data":"78f5af45de79ea61883a60b7c778c33a36e1bb9c47b7fd8e5b4ede94190cfcc1"} Mar 17 09:48:24 crc kubenswrapper[4813]: I0317 09:48:24.231759 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h6wdl" podStartSLOduration=2.695872383 podStartE2EDuration="5.231728295s" podCreationTimestamp="2026-03-17 09:48:19 +0000 UTC" firstStartedPulling="2026-03-17 09:48:21.169648994 +0000 UTC m=+2323.270452493" lastFinishedPulling="2026-03-17 09:48:23.705504866 +0000 UTC m=+2325.806308405" observedRunningTime="2026-03-17 09:48:24.22472848 +0000 UTC m=+2326.325532029" watchObservedRunningTime="2026-03-17 09:48:24.231728295 +0000 UTC m=+2326.332531834" Mar 17 09:48:29 crc kubenswrapper[4813]: I0317 09:48:29.730884 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:48:29 crc kubenswrapper[4813]: E0317 09:48:29.731664 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:48:30 crc kubenswrapper[4813]: I0317 09:48:30.392743 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h6wdl" Mar 17 09:48:30 crc kubenswrapper[4813]: I0317 09:48:30.392806 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h6wdl" Mar 17 09:48:30 crc kubenswrapper[4813]: I0317 09:48:30.449382 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h6wdl" Mar 17 09:48:31 crc kubenswrapper[4813]: I0317 09:48:31.323236 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h6wdl" Mar 17 09:48:32 crc kubenswrapper[4813]: I0317 09:48:32.175428 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h6wdl"] Mar 17 09:48:33 crc kubenswrapper[4813]: I0317 09:48:33.290798 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h6wdl" podUID="561032e6-7a20-40de-a141-88b07c2ee5ef" containerName="registry-server" containerID="cri-o://78f5af45de79ea61883a60b7c778c33a36e1bb9c47b7fd8e5b4ede94190cfcc1" gracePeriod=2 Mar 17 09:48:34 crc kubenswrapper[4813]: I0317 09:48:34.309088 4813 generic.go:334] "Generic (PLEG): container finished" podID="561032e6-7a20-40de-a141-88b07c2ee5ef" containerID="78f5af45de79ea61883a60b7c778c33a36e1bb9c47b7fd8e5b4ede94190cfcc1" exitCode=0 Mar 17 09:48:34 crc kubenswrapper[4813]: I0317 09:48:34.309145 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h6wdl" event={"ID":"561032e6-7a20-40de-a141-88b07c2ee5ef","Type":"ContainerDied","Data":"78f5af45de79ea61883a60b7c778c33a36e1bb9c47b7fd8e5b4ede94190cfcc1"} Mar 17 09:48:34 crc kubenswrapper[4813]: I0317 09:48:34.309577 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h6wdl" event={"ID":"561032e6-7a20-40de-a141-88b07c2ee5ef","Type":"ContainerDied","Data":"4a263e02580cbc1bd52cd87523865bc6d9d39c3d54645c30df23466f9bf8b76d"} Mar 17 09:48:34 crc kubenswrapper[4813]: I0317 09:48:34.309595 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a263e02580cbc1bd52cd87523865bc6d9d39c3d54645c30df23466f9bf8b76d" Mar 17 09:48:34 crc kubenswrapper[4813]: I0317 09:48:34.396890 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h6wdl" Mar 17 09:48:34 crc kubenswrapper[4813]: I0317 09:48:34.518426 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/561032e6-7a20-40de-a141-88b07c2ee5ef-catalog-content\") pod \"561032e6-7a20-40de-a141-88b07c2ee5ef\" (UID: \"561032e6-7a20-40de-a141-88b07c2ee5ef\") " Mar 17 09:48:34 crc kubenswrapper[4813]: I0317 09:48:34.518555 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k77dq\" (UniqueName: \"kubernetes.io/projected/561032e6-7a20-40de-a141-88b07c2ee5ef-kube-api-access-k77dq\") pod \"561032e6-7a20-40de-a141-88b07c2ee5ef\" (UID: \"561032e6-7a20-40de-a141-88b07c2ee5ef\") " Mar 17 09:48:34 crc kubenswrapper[4813]: I0317 09:48:34.518592 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/561032e6-7a20-40de-a141-88b07c2ee5ef-utilities\") pod \"561032e6-7a20-40de-a141-88b07c2ee5ef\" (UID: \"561032e6-7a20-40de-a141-88b07c2ee5ef\") " Mar 17 09:48:34 crc kubenswrapper[4813]: I0317 09:48:34.519549 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/561032e6-7a20-40de-a141-88b07c2ee5ef-utilities" (OuterVolumeSpecName: "utilities") pod "561032e6-7a20-40de-a141-88b07c2ee5ef" (UID: "561032e6-7a20-40de-a141-88b07c2ee5ef"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:48:34 crc kubenswrapper[4813]: I0317 09:48:34.533155 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/561032e6-7a20-40de-a141-88b07c2ee5ef-kube-api-access-k77dq" (OuterVolumeSpecName: "kube-api-access-k77dq") pod "561032e6-7a20-40de-a141-88b07c2ee5ef" (UID: "561032e6-7a20-40de-a141-88b07c2ee5ef"). InnerVolumeSpecName "kube-api-access-k77dq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:48:34 crc kubenswrapper[4813]: I0317 09:48:34.548085 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/561032e6-7a20-40de-a141-88b07c2ee5ef-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "561032e6-7a20-40de-a141-88b07c2ee5ef" (UID: "561032e6-7a20-40de-a141-88b07c2ee5ef"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:48:34 crc kubenswrapper[4813]: I0317 09:48:34.621262 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k77dq\" (UniqueName: \"kubernetes.io/projected/561032e6-7a20-40de-a141-88b07c2ee5ef-kube-api-access-k77dq\") on node \"crc\" DevicePath \"\"" Mar 17 09:48:34 crc kubenswrapper[4813]: I0317 09:48:34.621301 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/561032e6-7a20-40de-a141-88b07c2ee5ef-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:48:34 crc kubenswrapper[4813]: I0317 09:48:34.621311 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/561032e6-7a20-40de-a141-88b07c2ee5ef-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:48:35 crc kubenswrapper[4813]: I0317 09:48:35.320018 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h6wdl" Mar 17 09:48:35 crc kubenswrapper[4813]: I0317 09:48:35.355261 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h6wdl"] Mar 17 09:48:35 crc kubenswrapper[4813]: I0317 09:48:35.373512 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h6wdl"] Mar 17 09:48:36 crc kubenswrapper[4813]: I0317 09:48:36.744475 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="561032e6-7a20-40de-a141-88b07c2ee5ef" path="/var/lib/kubelet/pods/561032e6-7a20-40de-a141-88b07c2ee5ef/volumes" Mar 17 09:48:40 crc kubenswrapper[4813]: I0317 09:48:40.732118 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:48:40 crc kubenswrapper[4813]: E0317 09:48:40.733042 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:48:48 crc kubenswrapper[4813]: I0317 09:48:48.950800 4813 scope.go:117] "RemoveContainer" containerID="1513f492fdb771fc3f67a5a814465f90982e5571d9d36fcd2ca7bdc8fed3a084" Mar 17 09:48:52 crc kubenswrapper[4813]: I0317 09:48:52.733882 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:48:52 crc kubenswrapper[4813]: E0317 09:48:52.734725 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:49:05 crc kubenswrapper[4813]: I0317 09:49:05.731428 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:49:05 crc kubenswrapper[4813]: E0317 09:49:05.732563 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:49:17 crc kubenswrapper[4813]: I0317 09:49:17.730575 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:49:17 crc kubenswrapper[4813]: E0317 09:49:17.733889 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:49:31 crc kubenswrapper[4813]: I0317 09:49:31.731657 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:49:31 crc kubenswrapper[4813]: E0317 09:49:31.734234 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:49:42 crc kubenswrapper[4813]: I0317 09:49:42.731417 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:49:42 crc kubenswrapper[4813]: E0317 09:49:42.732692 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:49:54 crc kubenswrapper[4813]: I0317 09:49:54.731037 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:49:54 crc kubenswrapper[4813]: E0317 09:49:54.731833 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:50:00 crc kubenswrapper[4813]: I0317 09:50:00.155987 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562350-6hx9r"] Mar 17 09:50:00 crc kubenswrapper[4813]: E0317 09:50:00.156942 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561032e6-7a20-40de-a141-88b07c2ee5ef" containerName="extract-utilities" Mar 17 09:50:00 crc kubenswrapper[4813]: I0317 09:50:00.156956 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="561032e6-7a20-40de-a141-88b07c2ee5ef" containerName="extract-utilities" Mar 17 09:50:00 crc kubenswrapper[4813]: E0317 09:50:00.156977 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561032e6-7a20-40de-a141-88b07c2ee5ef" containerName="registry-server" Mar 17 09:50:00 crc kubenswrapper[4813]: I0317 09:50:00.156984 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="561032e6-7a20-40de-a141-88b07c2ee5ef" containerName="registry-server" Mar 17 09:50:00 crc kubenswrapper[4813]: E0317 09:50:00.157006 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561032e6-7a20-40de-a141-88b07c2ee5ef" containerName="extract-content" Mar 17 09:50:00 crc kubenswrapper[4813]: I0317 09:50:00.157012 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="561032e6-7a20-40de-a141-88b07c2ee5ef" containerName="extract-content" Mar 17 09:50:00 crc kubenswrapper[4813]: I0317 09:50:00.157183 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="561032e6-7a20-40de-a141-88b07c2ee5ef" containerName="registry-server" Mar 17 09:50:00 crc kubenswrapper[4813]: I0317 09:50:00.157831 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562350-6hx9r" Mar 17 09:50:00 crc kubenswrapper[4813]: I0317 09:50:00.160429 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:50:00 crc kubenswrapper[4813]: I0317 09:50:00.160867 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:50:00 crc kubenswrapper[4813]: I0317 09:50:00.161084 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:50:00 crc kubenswrapper[4813]: I0317 09:50:00.175337 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562350-6hx9r"] Mar 17 09:50:00 crc kubenswrapper[4813]: I0317 09:50:00.260279 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h88mz\" (UniqueName: \"kubernetes.io/projected/d125196b-08fb-4b13-9930-daf241996e49-kube-api-access-h88mz\") pod \"auto-csr-approver-29562350-6hx9r\" (UID: \"d125196b-08fb-4b13-9930-daf241996e49\") " pod="openshift-infra/auto-csr-approver-29562350-6hx9r" Mar 17 09:50:00 crc kubenswrapper[4813]: I0317 09:50:00.362427 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h88mz\" (UniqueName: \"kubernetes.io/projected/d125196b-08fb-4b13-9930-daf241996e49-kube-api-access-h88mz\") pod \"auto-csr-approver-29562350-6hx9r\" (UID: \"d125196b-08fb-4b13-9930-daf241996e49\") " pod="openshift-infra/auto-csr-approver-29562350-6hx9r" Mar 17 09:50:00 crc kubenswrapper[4813]: I0317 09:50:00.383338 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h88mz\" (UniqueName: \"kubernetes.io/projected/d125196b-08fb-4b13-9930-daf241996e49-kube-api-access-h88mz\") pod \"auto-csr-approver-29562350-6hx9r\" (UID: \"d125196b-08fb-4b13-9930-daf241996e49\") " pod="openshift-infra/auto-csr-approver-29562350-6hx9r" Mar 17 09:50:00 crc kubenswrapper[4813]: I0317 09:50:00.487039 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562350-6hx9r" Mar 17 09:50:01 crc kubenswrapper[4813]: I0317 09:50:01.038499 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562350-6hx9r"] Mar 17 09:50:01 crc kubenswrapper[4813]: W0317 09:50:01.041926 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd125196b_08fb_4b13_9930_daf241996e49.slice/crio-3987a36e963ac4968e06e9a78fc5fa78fdb086e47543546b68394265297fc581 WatchSource:0}: Error finding container 3987a36e963ac4968e06e9a78fc5fa78fdb086e47543546b68394265297fc581: Status 404 returned error can't find the container with id 3987a36e963ac4968e06e9a78fc5fa78fdb086e47543546b68394265297fc581 Mar 17 09:50:01 crc kubenswrapper[4813]: I0317 09:50:01.332091 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562350-6hx9r" event={"ID":"d125196b-08fb-4b13-9930-daf241996e49","Type":"ContainerStarted","Data":"3987a36e963ac4968e06e9a78fc5fa78fdb086e47543546b68394265297fc581"} Mar 17 09:50:03 crc kubenswrapper[4813]: I0317 09:50:03.354717 4813 generic.go:334] "Generic (PLEG): container finished" podID="d125196b-08fb-4b13-9930-daf241996e49" containerID="806dd70c01c034a744ce0d93e725c97fa79fc217ed0954e2d050eaf82a87e54e" exitCode=0 Mar 17 09:50:03 crc kubenswrapper[4813]: I0317 09:50:03.354829 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562350-6hx9r" event={"ID":"d125196b-08fb-4b13-9930-daf241996e49","Type":"ContainerDied","Data":"806dd70c01c034a744ce0d93e725c97fa79fc217ed0954e2d050eaf82a87e54e"} Mar 17 09:50:04 crc kubenswrapper[4813]: I0317 09:50:04.807106 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562350-6hx9r" Mar 17 09:50:04 crc kubenswrapper[4813]: I0317 09:50:04.858299 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h88mz\" (UniqueName: \"kubernetes.io/projected/d125196b-08fb-4b13-9930-daf241996e49-kube-api-access-h88mz\") pod \"d125196b-08fb-4b13-9930-daf241996e49\" (UID: \"d125196b-08fb-4b13-9930-daf241996e49\") " Mar 17 09:50:04 crc kubenswrapper[4813]: I0317 09:50:04.865068 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d125196b-08fb-4b13-9930-daf241996e49-kube-api-access-h88mz" (OuterVolumeSpecName: "kube-api-access-h88mz") pod "d125196b-08fb-4b13-9930-daf241996e49" (UID: "d125196b-08fb-4b13-9930-daf241996e49"). InnerVolumeSpecName "kube-api-access-h88mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:50:04 crc kubenswrapper[4813]: I0317 09:50:04.960228 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h88mz\" (UniqueName: \"kubernetes.io/projected/d125196b-08fb-4b13-9930-daf241996e49-kube-api-access-h88mz\") on node \"crc\" DevicePath \"\"" Mar 17 09:50:05 crc kubenswrapper[4813]: I0317 09:50:05.378913 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562350-6hx9r" event={"ID":"d125196b-08fb-4b13-9930-daf241996e49","Type":"ContainerDied","Data":"3987a36e963ac4968e06e9a78fc5fa78fdb086e47543546b68394265297fc581"} Mar 17 09:50:05 crc kubenswrapper[4813]: I0317 09:50:05.378973 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3987a36e963ac4968e06e9a78fc5fa78fdb086e47543546b68394265297fc581" Mar 17 09:50:05 crc kubenswrapper[4813]: I0317 09:50:05.378971 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562350-6hx9r" Mar 17 09:50:05 crc kubenswrapper[4813]: I0317 09:50:05.918461 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562344-6kk6q"] Mar 17 09:50:05 crc kubenswrapper[4813]: I0317 09:50:05.934801 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562344-6kk6q"] Mar 17 09:50:06 crc kubenswrapper[4813]: I0317 09:50:06.731376 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:50:06 crc kubenswrapper[4813]: E0317 09:50:06.731933 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:50:06 crc kubenswrapper[4813]: I0317 09:50:06.750284 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a530431d-cc3a-4735-aaac-a26a54acce1f" path="/var/lib/kubelet/pods/a530431d-cc3a-4735-aaac-a26a54acce1f/volumes" Mar 17 09:50:21 crc kubenswrapper[4813]: I0317 09:50:21.731447 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:50:21 crc kubenswrapper[4813]: E0317 09:50:21.732492 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:50:33 crc kubenswrapper[4813]: I0317 09:50:33.730619 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:50:33 crc kubenswrapper[4813]: E0317 09:50:33.731636 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:50:45 crc kubenswrapper[4813]: I0317 09:50:45.731101 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:50:45 crc kubenswrapper[4813]: E0317 09:50:45.732229 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:50:49 crc kubenswrapper[4813]: I0317 09:50:49.053745 4813 scope.go:117] "RemoveContainer" containerID="297b8b57c58c954c92e9132a175ce9254e8182b45460d2aebd5aeecf52b0eb7b" Mar 17 09:50:51 crc kubenswrapper[4813]: I0317 09:50:51.920765 4813 generic.go:334] "Generic (PLEG): container finished" podID="71e0074f-5ee2-4f59-9184-c23495021bfd" containerID="8f6c8123ad3b3e604a8fa61380731e0b2bc2e024140d7b7186c3dcae3b49177a" exitCode=0 Mar 17 09:50:51 crc kubenswrapper[4813]: I0317 09:50:51.921301 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" event={"ID":"71e0074f-5ee2-4f59-9184-c23495021bfd","Type":"ContainerDied","Data":"8f6c8123ad3b3e604a8fa61380731e0b2bc2e024140d7b7186c3dcae3b49177a"} Mar 17 09:50:53 crc kubenswrapper[4813]: I0317 09:50:53.432581 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" Mar 17 09:50:53 crc kubenswrapper[4813]: I0317 09:50:53.617860 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-libvirt-secret-0\") pod \"71e0074f-5ee2-4f59-9184-c23495021bfd\" (UID: \"71e0074f-5ee2-4f59-9184-c23495021bfd\") " Mar 17 09:50:53 crc kubenswrapper[4813]: I0317 09:50:53.617996 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxhnb\" (UniqueName: \"kubernetes.io/projected/71e0074f-5ee2-4f59-9184-c23495021bfd-kube-api-access-bxhnb\") pod \"71e0074f-5ee2-4f59-9184-c23495021bfd\" (UID: \"71e0074f-5ee2-4f59-9184-c23495021bfd\") " Mar 17 09:50:53 crc kubenswrapper[4813]: I0317 09:50:53.618054 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-inventory\") pod \"71e0074f-5ee2-4f59-9184-c23495021bfd\" (UID: \"71e0074f-5ee2-4f59-9184-c23495021bfd\") " Mar 17 09:50:53 crc kubenswrapper[4813]: I0317 09:50:53.618095 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-ssh-key-openstack-edpm-ipam\") pod \"71e0074f-5ee2-4f59-9184-c23495021bfd\" (UID: \"71e0074f-5ee2-4f59-9184-c23495021bfd\") " Mar 17 09:50:53 crc kubenswrapper[4813]: I0317 09:50:53.618966 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-libvirt-combined-ca-bundle\") pod \"71e0074f-5ee2-4f59-9184-c23495021bfd\" (UID: \"71e0074f-5ee2-4f59-9184-c23495021bfd\") " Mar 17 09:50:53 crc kubenswrapper[4813]: I0317 09:50:53.626709 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "71e0074f-5ee2-4f59-9184-c23495021bfd" (UID: "71e0074f-5ee2-4f59-9184-c23495021bfd"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:50:53 crc kubenswrapper[4813]: I0317 09:50:53.626870 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71e0074f-5ee2-4f59-9184-c23495021bfd-kube-api-access-bxhnb" (OuterVolumeSpecName: "kube-api-access-bxhnb") pod "71e0074f-5ee2-4f59-9184-c23495021bfd" (UID: "71e0074f-5ee2-4f59-9184-c23495021bfd"). InnerVolumeSpecName "kube-api-access-bxhnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:50:53 crc kubenswrapper[4813]: I0317 09:50:53.659937 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-inventory" (OuterVolumeSpecName: "inventory") pod "71e0074f-5ee2-4f59-9184-c23495021bfd" (UID: "71e0074f-5ee2-4f59-9184-c23495021bfd"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:50:53 crc kubenswrapper[4813]: I0317 09:50:53.666907 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "71e0074f-5ee2-4f59-9184-c23495021bfd" (UID: "71e0074f-5ee2-4f59-9184-c23495021bfd"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:50:53 crc kubenswrapper[4813]: I0317 09:50:53.672579 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "71e0074f-5ee2-4f59-9184-c23495021bfd" (UID: "71e0074f-5ee2-4f59-9184-c23495021bfd"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:50:53 crc kubenswrapper[4813]: I0317 09:50:53.721905 4813 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:50:53 crc kubenswrapper[4813]: I0317 09:50:53.721955 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxhnb\" (UniqueName: \"kubernetes.io/projected/71e0074f-5ee2-4f59-9184-c23495021bfd-kube-api-access-bxhnb\") on node \"crc\" DevicePath \"\"" Mar 17 09:50:53 crc kubenswrapper[4813]: I0317 09:50:53.721977 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 09:50:53 crc kubenswrapper[4813]: I0317 09:50:53.722000 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 09:50:53 crc kubenswrapper[4813]: I0317 09:50:53.722020 4813 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71e0074f-5ee2-4f59-9184-c23495021bfd-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:50:53 crc kubenswrapper[4813]: I0317 09:50:53.948836 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" event={"ID":"71e0074f-5ee2-4f59-9184-c23495021bfd","Type":"ContainerDied","Data":"befc1780e63f4560aec113dc77461d850713071f980743276dc5aa119972c97e"} Mar 17 09:50:53 crc kubenswrapper[4813]: I0317 09:50:53.948899 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="befc1780e63f4560aec113dc77461d850713071f980743276dc5aa119972c97e" Mar 17 09:50:53 crc kubenswrapper[4813]: I0317 09:50:53.948975 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.141151 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk"] Mar 17 09:50:54 crc kubenswrapper[4813]: E0317 09:50:54.141759 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71e0074f-5ee2-4f59-9184-c23495021bfd" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.141790 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="71e0074f-5ee2-4f59-9184-c23495021bfd" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 17 09:50:54 crc kubenswrapper[4813]: E0317 09:50:54.141840 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d125196b-08fb-4b13-9930-daf241996e49" containerName="oc" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.141853 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d125196b-08fb-4b13-9930-daf241996e49" containerName="oc" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.142122 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="71e0074f-5ee2-4f59-9184-c23495021bfd" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.142152 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d125196b-08fb-4b13-9930-daf241996e49" containerName="oc" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.143139 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.146240 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.146735 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.148396 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.148420 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.148471 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.148473 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.149065 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hj9h4" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.182438 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk"] Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.234579 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.234663 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.234700 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.234780 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.234847 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.234902 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.234980 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/bd00eea2-860c-40ac-8147-41992af388ec-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.235154 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.235191 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.235221 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqp6c\" (UniqueName: \"kubernetes.io/projected/bd00eea2-860c-40ac-8147-41992af388ec-kube-api-access-xqp6c\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.235334 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.336885 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.337021 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.337080 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.337135 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.337224 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.337265 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.337320 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.337370 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/bd00eea2-860c-40ac-8147-41992af388ec-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.337460 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.337510 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.337561 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqp6c\" (UniqueName: \"kubernetes.io/projected/bd00eea2-860c-40ac-8147-41992af388ec-kube-api-access-xqp6c\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.338793 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/bd00eea2-860c-40ac-8147-41992af388ec-nova-extra-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.341086 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-inventory\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.341318 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-3\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.341368 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-2\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.341571 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-ssh-key-openstack-edpm-ipam\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.342630 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-migration-ssh-key-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.343984 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.344111 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-0\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.353204 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-combined-ca-bundle\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.355245 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-migration-ssh-key-1\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.355642 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqp6c\" (UniqueName: \"kubernetes.io/projected/bd00eea2-860c-40ac-8147-41992af388ec-kube-api-access-xqp6c\") pod \"nova-edpm-deployment-openstack-edpm-ipam-6z5zk\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:54 crc kubenswrapper[4813]: I0317 09:50:54.477566 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:50:55 crc kubenswrapper[4813]: I0317 09:50:55.133008 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk"] Mar 17 09:50:55 crc kubenswrapper[4813]: I0317 09:50:55.981622 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" event={"ID":"bd00eea2-860c-40ac-8147-41992af388ec","Type":"ContainerStarted","Data":"233e074a952480de256bc0627141a297f27cbe5ef4b4661788254d27e4829e79"} Mar 17 09:50:56 crc kubenswrapper[4813]: I0317 09:50:56.997685 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" event={"ID":"bd00eea2-860c-40ac-8147-41992af388ec","Type":"ContainerStarted","Data":"77827f6ea3da52dda5319780506c792f6a7b882e8ab942b1b06661b319bb584a"} Mar 17 09:50:57 crc kubenswrapper[4813]: I0317 09:50:57.027736 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" podStartSLOduration=2.49857761 podStartE2EDuration="3.027715241s" podCreationTimestamp="2026-03-17 09:50:54 +0000 UTC" firstStartedPulling="2026-03-17 09:50:55.149423644 +0000 UTC m=+2477.250227183" lastFinishedPulling="2026-03-17 09:50:55.678561285 +0000 UTC m=+2477.779364814" observedRunningTime="2026-03-17 09:50:57.019848197 +0000 UTC m=+2479.120651726" watchObservedRunningTime="2026-03-17 09:50:57.027715241 +0000 UTC m=+2479.128518750" Mar 17 09:50:57 crc kubenswrapper[4813]: I0317 09:50:57.731231 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:50:57 crc kubenswrapper[4813]: E0317 09:50:57.731873 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:51:11 crc kubenswrapper[4813]: I0317 09:51:11.730932 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:51:11 crc kubenswrapper[4813]: E0317 09:51:11.732466 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:51:24 crc kubenswrapper[4813]: I0317 09:51:24.736319 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:51:24 crc kubenswrapper[4813]: E0317 09:51:24.737726 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:51:37 crc kubenswrapper[4813]: I0317 09:51:37.731997 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:51:37 crc kubenswrapper[4813]: E0317 09:51:37.733074 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:51:48 crc kubenswrapper[4813]: I0317 09:51:48.743583 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:51:48 crc kubenswrapper[4813]: E0317 09:51:48.744410 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:52:00 crc kubenswrapper[4813]: I0317 09:52:00.168409 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562352-n4nx6"] Mar 17 09:52:00 crc kubenswrapper[4813]: I0317 09:52:00.172523 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562352-n4nx6" Mar 17 09:52:00 crc kubenswrapper[4813]: I0317 09:52:00.175857 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:52:00 crc kubenswrapper[4813]: I0317 09:52:00.176338 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:52:00 crc kubenswrapper[4813]: I0317 09:52:00.176370 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:52:00 crc kubenswrapper[4813]: I0317 09:52:00.182178 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562352-n4nx6"] Mar 17 09:52:00 crc kubenswrapper[4813]: I0317 09:52:00.230569 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktxgc\" (UniqueName: \"kubernetes.io/projected/0cb5aa99-7a25-494a-b945-925a88a3f5d7-kube-api-access-ktxgc\") pod \"auto-csr-approver-29562352-n4nx6\" (UID: \"0cb5aa99-7a25-494a-b945-925a88a3f5d7\") " pod="openshift-infra/auto-csr-approver-29562352-n4nx6" Mar 17 09:52:00 crc kubenswrapper[4813]: I0317 09:52:00.332654 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktxgc\" (UniqueName: \"kubernetes.io/projected/0cb5aa99-7a25-494a-b945-925a88a3f5d7-kube-api-access-ktxgc\") pod \"auto-csr-approver-29562352-n4nx6\" (UID: \"0cb5aa99-7a25-494a-b945-925a88a3f5d7\") " pod="openshift-infra/auto-csr-approver-29562352-n4nx6" Mar 17 09:52:00 crc kubenswrapper[4813]: I0317 09:52:00.377558 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktxgc\" (UniqueName: \"kubernetes.io/projected/0cb5aa99-7a25-494a-b945-925a88a3f5d7-kube-api-access-ktxgc\") pod \"auto-csr-approver-29562352-n4nx6\" (UID: \"0cb5aa99-7a25-494a-b945-925a88a3f5d7\") " pod="openshift-infra/auto-csr-approver-29562352-n4nx6" Mar 17 09:52:00 crc kubenswrapper[4813]: I0317 09:52:00.501885 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562352-n4nx6" Mar 17 09:52:00 crc kubenswrapper[4813]: I0317 09:52:00.844692 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562352-n4nx6"] Mar 17 09:52:00 crc kubenswrapper[4813]: I0317 09:52:00.859208 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 09:52:01 crc kubenswrapper[4813]: I0317 09:52:01.754502 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562352-n4nx6" event={"ID":"0cb5aa99-7a25-494a-b945-925a88a3f5d7","Type":"ContainerStarted","Data":"d3b86536331d2c94a3ac15b548d5b0eca53243ece95551f029c59f0a498571de"} Mar 17 09:52:02 crc kubenswrapper[4813]: I0317 09:52:02.765876 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562352-n4nx6" event={"ID":"0cb5aa99-7a25-494a-b945-925a88a3f5d7","Type":"ContainerStarted","Data":"8c153ea76086c7849bf876e64e764357f065dace947978f32afb995852f805e0"} Mar 17 09:52:02 crc kubenswrapper[4813]: I0317 09:52:02.804593 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562352-n4nx6" podStartSLOduration=1.387297016 podStartE2EDuration="2.80456709s" podCreationTimestamp="2026-03-17 09:52:00 +0000 UTC" firstStartedPulling="2026-03-17 09:52:00.858897358 +0000 UTC m=+2542.959700867" lastFinishedPulling="2026-03-17 09:52:02.276167422 +0000 UTC m=+2544.376970941" observedRunningTime="2026-03-17 09:52:02.784747536 +0000 UTC m=+2544.885551065" watchObservedRunningTime="2026-03-17 09:52:02.80456709 +0000 UTC m=+2544.905370619" Mar 17 09:52:03 crc kubenswrapper[4813]: I0317 09:52:03.731489 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:52:03 crc kubenswrapper[4813]: E0317 09:52:03.731739 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:52:03 crc kubenswrapper[4813]: I0317 09:52:03.779685 4813 generic.go:334] "Generic (PLEG): container finished" podID="0cb5aa99-7a25-494a-b945-925a88a3f5d7" containerID="8c153ea76086c7849bf876e64e764357f065dace947978f32afb995852f805e0" exitCode=0 Mar 17 09:52:03 crc kubenswrapper[4813]: I0317 09:52:03.779744 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562352-n4nx6" event={"ID":"0cb5aa99-7a25-494a-b945-925a88a3f5d7","Type":"ContainerDied","Data":"8c153ea76086c7849bf876e64e764357f065dace947978f32afb995852f805e0"} Mar 17 09:52:05 crc kubenswrapper[4813]: I0317 09:52:05.153885 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562352-n4nx6" Mar 17 09:52:05 crc kubenswrapper[4813]: I0317 09:52:05.249735 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ktxgc\" (UniqueName: \"kubernetes.io/projected/0cb5aa99-7a25-494a-b945-925a88a3f5d7-kube-api-access-ktxgc\") pod \"0cb5aa99-7a25-494a-b945-925a88a3f5d7\" (UID: \"0cb5aa99-7a25-494a-b945-925a88a3f5d7\") " Mar 17 09:52:05 crc kubenswrapper[4813]: I0317 09:52:05.258744 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0cb5aa99-7a25-494a-b945-925a88a3f5d7-kube-api-access-ktxgc" (OuterVolumeSpecName: "kube-api-access-ktxgc") pod "0cb5aa99-7a25-494a-b945-925a88a3f5d7" (UID: "0cb5aa99-7a25-494a-b945-925a88a3f5d7"). InnerVolumeSpecName "kube-api-access-ktxgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:52:05 crc kubenswrapper[4813]: I0317 09:52:05.352669 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ktxgc\" (UniqueName: \"kubernetes.io/projected/0cb5aa99-7a25-494a-b945-925a88a3f5d7-kube-api-access-ktxgc\") on node \"crc\" DevicePath \"\"" Mar 17 09:52:05 crc kubenswrapper[4813]: I0317 09:52:05.802267 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562352-n4nx6" event={"ID":"0cb5aa99-7a25-494a-b945-925a88a3f5d7","Type":"ContainerDied","Data":"d3b86536331d2c94a3ac15b548d5b0eca53243ece95551f029c59f0a498571de"} Mar 17 09:52:05 crc kubenswrapper[4813]: I0317 09:52:05.802332 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d3b86536331d2c94a3ac15b548d5b0eca53243ece95551f029c59f0a498571de" Mar 17 09:52:05 crc kubenswrapper[4813]: I0317 09:52:05.802374 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562352-n4nx6" Mar 17 09:52:05 crc kubenswrapper[4813]: I0317 09:52:05.865093 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562346-rrzcl"] Mar 17 09:52:05 crc kubenswrapper[4813]: I0317 09:52:05.872734 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562346-rrzcl"] Mar 17 09:52:06 crc kubenswrapper[4813]: I0317 09:52:06.750440 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4545fea0-ba23-4b1a-9640-bec6c63070dd" path="/var/lib/kubelet/pods/4545fea0-ba23-4b1a-9640-bec6c63070dd/volumes" Mar 17 09:52:16 crc kubenswrapper[4813]: I0317 09:52:16.731513 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:52:16 crc kubenswrapper[4813]: E0317 09:52:16.732875 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:52:28 crc kubenswrapper[4813]: I0317 09:52:28.742983 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:52:28 crc kubenswrapper[4813]: E0317 09:52:28.744724 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:52:39 crc kubenswrapper[4813]: I0317 09:52:39.731031 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:52:39 crc kubenswrapper[4813]: E0317 09:52:39.732302 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:52:49 crc kubenswrapper[4813]: I0317 09:52:49.162538 4813 scope.go:117] "RemoveContainer" containerID="5f6299aaaf5f21de7750f72626a9708027e7ede92ce6871882347c5aa616330f" Mar 17 09:52:50 crc kubenswrapper[4813]: I0317 09:52:50.730927 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:52:50 crc kubenswrapper[4813]: E0317 09:52:50.731522 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:53:05 crc kubenswrapper[4813]: I0317 09:53:05.731972 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:53:05 crc kubenswrapper[4813]: E0317 09:53:05.733021 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:53:17 crc kubenswrapper[4813]: I0317 09:53:17.730884 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:53:18 crc kubenswrapper[4813]: I0317 09:53:18.656043 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerStarted","Data":"df44ca5fc960279ac359c596c0c46ba2f6c33c03061a11631b7e3d791130963b"} Mar 17 09:53:32 crc kubenswrapper[4813]: I0317 09:53:32.805154 4813 generic.go:334] "Generic (PLEG): container finished" podID="bd00eea2-860c-40ac-8147-41992af388ec" containerID="77827f6ea3da52dda5319780506c792f6a7b882e8ab942b1b06661b319bb584a" exitCode=0 Mar 17 09:53:32 crc kubenswrapper[4813]: I0317 09:53:32.805258 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" event={"ID":"bd00eea2-860c-40ac-8147-41992af388ec","Type":"ContainerDied","Data":"77827f6ea3da52dda5319780506c792f6a7b882e8ab942b1b06661b319bb584a"} Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.344703 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.538338 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqp6c\" (UniqueName: \"kubernetes.io/projected/bd00eea2-860c-40ac-8147-41992af388ec-kube-api-access-xqp6c\") pod \"bd00eea2-860c-40ac-8147-41992af388ec\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.538434 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-3\") pod \"bd00eea2-860c-40ac-8147-41992af388ec\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.538499 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-inventory\") pod \"bd00eea2-860c-40ac-8147-41992af388ec\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.538558 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/bd00eea2-860c-40ac-8147-41992af388ec-nova-extra-config-0\") pod \"bd00eea2-860c-40ac-8147-41992af388ec\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.538640 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-0\") pod \"bd00eea2-860c-40ac-8147-41992af388ec\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.538748 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-ssh-key-openstack-edpm-ipam\") pod \"bd00eea2-860c-40ac-8147-41992af388ec\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.538868 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-1\") pod \"bd00eea2-860c-40ac-8147-41992af388ec\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.538955 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-migration-ssh-key-1\") pod \"bd00eea2-860c-40ac-8147-41992af388ec\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.539009 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-2\") pod \"bd00eea2-860c-40ac-8147-41992af388ec\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.539140 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-combined-ca-bundle\") pod \"bd00eea2-860c-40ac-8147-41992af388ec\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.539194 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-migration-ssh-key-0\") pod \"bd00eea2-860c-40ac-8147-41992af388ec\" (UID: \"bd00eea2-860c-40ac-8147-41992af388ec\") " Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.560990 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "bd00eea2-860c-40ac-8147-41992af388ec" (UID: "bd00eea2-860c-40ac-8147-41992af388ec"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.571166 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-2" (OuterVolumeSpecName: "nova-cell1-compute-config-2") pod "bd00eea2-860c-40ac-8147-41992af388ec" (UID: "bd00eea2-860c-40ac-8147-41992af388ec"). InnerVolumeSpecName "nova-cell1-compute-config-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.573658 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd00eea2-860c-40ac-8147-41992af388ec-kube-api-access-xqp6c" (OuterVolumeSpecName: "kube-api-access-xqp6c") pod "bd00eea2-860c-40ac-8147-41992af388ec" (UID: "bd00eea2-860c-40ac-8147-41992af388ec"). InnerVolumeSpecName "kube-api-access-xqp6c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.587990 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "bd00eea2-860c-40ac-8147-41992af388ec" (UID: "bd00eea2-860c-40ac-8147-41992af388ec"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.588772 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-inventory" (OuterVolumeSpecName: "inventory") pod "bd00eea2-860c-40ac-8147-41992af388ec" (UID: "bd00eea2-860c-40ac-8147-41992af388ec"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.589343 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "bd00eea2-860c-40ac-8147-41992af388ec" (UID: "bd00eea2-860c-40ac-8147-41992af388ec"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.600496 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "bd00eea2-860c-40ac-8147-41992af388ec" (UID: "bd00eea2-860c-40ac-8147-41992af388ec"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.603744 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "bd00eea2-860c-40ac-8147-41992af388ec" (UID: "bd00eea2-860c-40ac-8147-41992af388ec"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.605769 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "bd00eea2-860c-40ac-8147-41992af388ec" (UID: "bd00eea2-860c-40ac-8147-41992af388ec"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.617695 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd00eea2-860c-40ac-8147-41992af388ec-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "bd00eea2-860c-40ac-8147-41992af388ec" (UID: "bd00eea2-860c-40ac-8147-41992af388ec"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.620993 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-3" (OuterVolumeSpecName: "nova-cell1-compute-config-3") pod "bd00eea2-860c-40ac-8147-41992af388ec" (UID: "bd00eea2-860c-40ac-8147-41992af388ec"). InnerVolumeSpecName "nova-cell1-compute-config-3". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.642326 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.642514 4813 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.642613 4813 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.642772 4813 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-2\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-2\") on node \"crc\" DevicePath \"\"" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.642937 4813 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.642997 4813 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.643094 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqp6c\" (UniqueName: \"kubernetes.io/projected/bd00eea2-860c-40ac-8147-41992af388ec-kube-api-access-xqp6c\") on node \"crc\" DevicePath \"\"" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.643255 4813 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-3\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-3\") on node \"crc\" DevicePath \"\"" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.643422 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.643529 4813 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/bd00eea2-860c-40ac-8147-41992af388ec-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.643696 4813 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/bd00eea2-860c-40ac-8147-41992af388ec-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.828513 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" event={"ID":"bd00eea2-860c-40ac-8147-41992af388ec","Type":"ContainerDied","Data":"233e074a952480de256bc0627141a297f27cbe5ef4b4661788254d27e4829e79"} Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.828872 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="233e074a952480de256bc0627141a297f27cbe5ef4b4661788254d27e4829e79" Mar 17 09:53:34 crc kubenswrapper[4813]: I0317 09:53:34.828585 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-edpm-deployment-openstack-edpm-ipam-6z5zk" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.019103 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp"] Mar 17 09:53:35 crc kubenswrapper[4813]: E0317 09:53:35.019460 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd00eea2-860c-40ac-8147-41992af388ec" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.019472 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd00eea2-860c-40ac-8147-41992af388ec" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 17 09:53:35 crc kubenswrapper[4813]: E0317 09:53:35.019508 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0cb5aa99-7a25-494a-b945-925a88a3f5d7" containerName="oc" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.019517 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0cb5aa99-7a25-494a-b945-925a88a3f5d7" containerName="oc" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.019753 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0cb5aa99-7a25-494a-b945-925a88a3f5d7" containerName="oc" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.019768 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd00eea2-860c-40ac-8147-41992af388ec" containerName="nova-edpm-deployment-openstack-edpm-ipam" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.020431 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.022759 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.024024 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.024071 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.024308 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-hj9h4" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.024846 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.051011 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp"] Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.155592 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-972kp\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.155811 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-972kp\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.155908 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lt7s2\" (UniqueName: \"kubernetes.io/projected/cd1a58de-c5c8-4b15-a894-47baac35d6e2-kube-api-access-lt7s2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-972kp\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.156108 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-972kp\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.156170 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-972kp\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.156414 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-972kp\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.156522 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-972kp\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.258786 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-972kp\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.258844 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lt7s2\" (UniqueName: \"kubernetes.io/projected/cd1a58de-c5c8-4b15-a894-47baac35d6e2-kube-api-access-lt7s2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-972kp\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.258920 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-972kp\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.258943 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-972kp\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.259028 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-972kp\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.259062 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-972kp\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.259141 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-972kp\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.267645 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-972kp\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.268140 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-972kp\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.268131 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ssh-key-openstack-edpm-ipam\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-972kp\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.268372 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-972kp\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.284666 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lt7s2\" (UniqueName: \"kubernetes.io/projected/cd1a58de-c5c8-4b15-a894-47baac35d6e2-kube-api-access-lt7s2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-972kp\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.284924 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-972kp\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.296110 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-972kp\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:35 crc kubenswrapper[4813]: I0317 09:53:35.342306 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:53:36 crc kubenswrapper[4813]: I0317 09:53:36.003564 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp"] Mar 17 09:53:36 crc kubenswrapper[4813]: I0317 09:53:36.856294 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" event={"ID":"cd1a58de-c5c8-4b15-a894-47baac35d6e2","Type":"ContainerStarted","Data":"afd527dab194ddc87ce9959e18af104a344ca53b33c34a66310db549bf3bbc76"} Mar 17 09:53:36 crc kubenswrapper[4813]: I0317 09:53:36.857786 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" event={"ID":"cd1a58de-c5c8-4b15-a894-47baac35d6e2","Type":"ContainerStarted","Data":"64a84a76c5d479fde74df9186af00359d6a78efad02a1554d95353e8b8e2d239"} Mar 17 09:53:36 crc kubenswrapper[4813]: I0317 09:53:36.885445 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" podStartSLOduration=2.464391697 podStartE2EDuration="2.885424321s" podCreationTimestamp="2026-03-17 09:53:34 +0000 UTC" firstStartedPulling="2026-03-17 09:53:36.014739617 +0000 UTC m=+2638.115543126" lastFinishedPulling="2026-03-17 09:53:36.435772251 +0000 UTC m=+2638.536575750" observedRunningTime="2026-03-17 09:53:36.878359313 +0000 UTC m=+2638.979162852" watchObservedRunningTime="2026-03-17 09:53:36.885424321 +0000 UTC m=+2638.986227820" Mar 17 09:54:00 crc kubenswrapper[4813]: I0317 09:54:00.152960 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562354-28hkn"] Mar 17 09:54:00 crc kubenswrapper[4813]: I0317 09:54:00.155322 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562354-28hkn" Mar 17 09:54:00 crc kubenswrapper[4813]: I0317 09:54:00.162406 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:54:00 crc kubenswrapper[4813]: I0317 09:54:00.162985 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:54:00 crc kubenswrapper[4813]: I0317 09:54:00.163245 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:54:00 crc kubenswrapper[4813]: I0317 09:54:00.168251 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562354-28hkn"] Mar 17 09:54:00 crc kubenswrapper[4813]: I0317 09:54:00.253267 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbjfz\" (UniqueName: \"kubernetes.io/projected/b5b02c90-782e-4dd8-81e7-44e868a1eb64-kube-api-access-cbjfz\") pod \"auto-csr-approver-29562354-28hkn\" (UID: \"b5b02c90-782e-4dd8-81e7-44e868a1eb64\") " pod="openshift-infra/auto-csr-approver-29562354-28hkn" Mar 17 09:54:00 crc kubenswrapper[4813]: I0317 09:54:00.355140 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbjfz\" (UniqueName: \"kubernetes.io/projected/b5b02c90-782e-4dd8-81e7-44e868a1eb64-kube-api-access-cbjfz\") pod \"auto-csr-approver-29562354-28hkn\" (UID: \"b5b02c90-782e-4dd8-81e7-44e868a1eb64\") " pod="openshift-infra/auto-csr-approver-29562354-28hkn" Mar 17 09:54:00 crc kubenswrapper[4813]: I0317 09:54:00.374138 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbjfz\" (UniqueName: \"kubernetes.io/projected/b5b02c90-782e-4dd8-81e7-44e868a1eb64-kube-api-access-cbjfz\") pod \"auto-csr-approver-29562354-28hkn\" (UID: \"b5b02c90-782e-4dd8-81e7-44e868a1eb64\") " pod="openshift-infra/auto-csr-approver-29562354-28hkn" Mar 17 09:54:00 crc kubenswrapper[4813]: I0317 09:54:00.499266 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562354-28hkn" Mar 17 09:54:00 crc kubenswrapper[4813]: I0317 09:54:00.957921 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562354-28hkn"] Mar 17 09:54:01 crc kubenswrapper[4813]: I0317 09:54:01.264523 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562354-28hkn" event={"ID":"b5b02c90-782e-4dd8-81e7-44e868a1eb64","Type":"ContainerStarted","Data":"75a63d1db5abc770e2a9e649ccc2b70bfa807654a00c416050d66f9d5fdf84c2"} Mar 17 09:54:02 crc kubenswrapper[4813]: I0317 09:54:02.291347 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562354-28hkn" podStartSLOduration=1.336446623 podStartE2EDuration="2.291326764s" podCreationTimestamp="2026-03-17 09:54:00 +0000 UTC" firstStartedPulling="2026-03-17 09:54:00.959206394 +0000 UTC m=+2663.060009893" lastFinishedPulling="2026-03-17 09:54:01.914086505 +0000 UTC m=+2664.014890034" observedRunningTime="2026-03-17 09:54:02.290807978 +0000 UTC m=+2664.391611507" watchObservedRunningTime="2026-03-17 09:54:02.291326764 +0000 UTC m=+2664.392130283" Mar 17 09:54:03 crc kubenswrapper[4813]: I0317 09:54:03.284024 4813 generic.go:334] "Generic (PLEG): container finished" podID="b5b02c90-782e-4dd8-81e7-44e868a1eb64" containerID="5460beadf7cd3ead87ea493ab4e4d67b338c964a2b007e8b646f9f96551be425" exitCode=0 Mar 17 09:54:03 crc kubenswrapper[4813]: I0317 09:54:03.284145 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562354-28hkn" event={"ID":"b5b02c90-782e-4dd8-81e7-44e868a1eb64","Type":"ContainerDied","Data":"5460beadf7cd3ead87ea493ab4e4d67b338c964a2b007e8b646f9f96551be425"} Mar 17 09:54:04 crc kubenswrapper[4813]: I0317 09:54:04.721554 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562354-28hkn" Mar 17 09:54:04 crc kubenswrapper[4813]: I0317 09:54:04.850825 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbjfz\" (UniqueName: \"kubernetes.io/projected/b5b02c90-782e-4dd8-81e7-44e868a1eb64-kube-api-access-cbjfz\") pod \"b5b02c90-782e-4dd8-81e7-44e868a1eb64\" (UID: \"b5b02c90-782e-4dd8-81e7-44e868a1eb64\") " Mar 17 09:54:04 crc kubenswrapper[4813]: I0317 09:54:04.856354 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5b02c90-782e-4dd8-81e7-44e868a1eb64-kube-api-access-cbjfz" (OuterVolumeSpecName: "kube-api-access-cbjfz") pod "b5b02c90-782e-4dd8-81e7-44e868a1eb64" (UID: "b5b02c90-782e-4dd8-81e7-44e868a1eb64"). InnerVolumeSpecName "kube-api-access-cbjfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:54:04 crc kubenswrapper[4813]: I0317 09:54:04.953006 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbjfz\" (UniqueName: \"kubernetes.io/projected/b5b02c90-782e-4dd8-81e7-44e868a1eb64-kube-api-access-cbjfz\") on node \"crc\" DevicePath \"\"" Mar 17 09:54:05 crc kubenswrapper[4813]: I0317 09:54:05.321069 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562354-28hkn" event={"ID":"b5b02c90-782e-4dd8-81e7-44e868a1eb64","Type":"ContainerDied","Data":"75a63d1db5abc770e2a9e649ccc2b70bfa807654a00c416050d66f9d5fdf84c2"} Mar 17 09:54:05 crc kubenswrapper[4813]: I0317 09:54:05.321144 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="75a63d1db5abc770e2a9e649ccc2b70bfa807654a00c416050d66f9d5fdf84c2" Mar 17 09:54:05 crc kubenswrapper[4813]: I0317 09:54:05.321249 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562354-28hkn" Mar 17 09:54:05 crc kubenswrapper[4813]: I0317 09:54:05.388648 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562348-lt7kt"] Mar 17 09:54:05 crc kubenswrapper[4813]: I0317 09:54:05.398333 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562348-lt7kt"] Mar 17 09:54:06 crc kubenswrapper[4813]: I0317 09:54:06.748356 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ed42eba9-9fb3-4505-acd2-c552b16139b0" path="/var/lib/kubelet/pods/ed42eba9-9fb3-4505-acd2-c552b16139b0/volumes" Mar 17 09:54:49 crc kubenswrapper[4813]: I0317 09:54:49.294243 4813 scope.go:117] "RemoveContainer" containerID="93bbbd082fb016f1f3994a86f419e32766d9241ae7608f82c19120e91e144157" Mar 17 09:54:49 crc kubenswrapper[4813]: I0317 09:54:49.354881 4813 scope.go:117] "RemoveContainer" containerID="0447f6b3327a1f75bcc942832fcb55fa3331bd34e952778f74c9d7754d62b14e" Mar 17 09:54:49 crc kubenswrapper[4813]: I0317 09:54:49.391231 4813 scope.go:117] "RemoveContainer" containerID="78f5af45de79ea61883a60b7c778c33a36e1bb9c47b7fd8e5b4ede94190cfcc1" Mar 17 09:54:49 crc kubenswrapper[4813]: I0317 09:54:49.439833 4813 scope.go:117] "RemoveContainer" containerID="f6b111c68dd60f004d5b9306150c269e2dcd53446715f73252dacd88b33c1a0b" Mar 17 09:55:44 crc kubenswrapper[4813]: I0317 09:55:44.114411 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:55:44 crc kubenswrapper[4813]: I0317 09:55:44.115245 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:56:00 crc kubenswrapper[4813]: I0317 09:56:00.152683 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562356-kdgd8"] Mar 17 09:56:00 crc kubenswrapper[4813]: E0317 09:56:00.153831 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5b02c90-782e-4dd8-81e7-44e868a1eb64" containerName="oc" Mar 17 09:56:00 crc kubenswrapper[4813]: I0317 09:56:00.153852 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5b02c90-782e-4dd8-81e7-44e868a1eb64" containerName="oc" Mar 17 09:56:00 crc kubenswrapper[4813]: I0317 09:56:00.154159 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5b02c90-782e-4dd8-81e7-44e868a1eb64" containerName="oc" Mar 17 09:56:00 crc kubenswrapper[4813]: I0317 09:56:00.155241 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562356-kdgd8" Mar 17 09:56:00 crc kubenswrapper[4813]: I0317 09:56:00.159030 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:56:00 crc kubenswrapper[4813]: I0317 09:56:00.159187 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:56:00 crc kubenswrapper[4813]: I0317 09:56:00.159487 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:56:00 crc kubenswrapper[4813]: I0317 09:56:00.173838 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562356-kdgd8"] Mar 17 09:56:00 crc kubenswrapper[4813]: I0317 09:56:00.346698 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4bvh\" (UniqueName: \"kubernetes.io/projected/069b5d80-8eea-46f2-b7e2-1843d75656d3-kube-api-access-g4bvh\") pod \"auto-csr-approver-29562356-kdgd8\" (UID: \"069b5d80-8eea-46f2-b7e2-1843d75656d3\") " pod="openshift-infra/auto-csr-approver-29562356-kdgd8" Mar 17 09:56:00 crc kubenswrapper[4813]: I0317 09:56:00.447983 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4bvh\" (UniqueName: \"kubernetes.io/projected/069b5d80-8eea-46f2-b7e2-1843d75656d3-kube-api-access-g4bvh\") pod \"auto-csr-approver-29562356-kdgd8\" (UID: \"069b5d80-8eea-46f2-b7e2-1843d75656d3\") " pod="openshift-infra/auto-csr-approver-29562356-kdgd8" Mar 17 09:56:00 crc kubenswrapper[4813]: I0317 09:56:00.472309 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4bvh\" (UniqueName: \"kubernetes.io/projected/069b5d80-8eea-46f2-b7e2-1843d75656d3-kube-api-access-g4bvh\") pod \"auto-csr-approver-29562356-kdgd8\" (UID: \"069b5d80-8eea-46f2-b7e2-1843d75656d3\") " pod="openshift-infra/auto-csr-approver-29562356-kdgd8" Mar 17 09:56:00 crc kubenswrapper[4813]: I0317 09:56:00.483873 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562356-kdgd8" Mar 17 09:56:01 crc kubenswrapper[4813]: I0317 09:56:01.498302 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562356-kdgd8"] Mar 17 09:56:01 crc kubenswrapper[4813]: W0317 09:56:01.507750 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod069b5d80_8eea_46f2_b7e2_1843d75656d3.slice/crio-3c6beb0903e9cbf86086e59294c2f7965cb0717b4db6da5907078a267282ded1 WatchSource:0}: Error finding container 3c6beb0903e9cbf86086e59294c2f7965cb0717b4db6da5907078a267282ded1: Status 404 returned error can't find the container with id 3c6beb0903e9cbf86086e59294c2f7965cb0717b4db6da5907078a267282ded1 Mar 17 09:56:02 crc kubenswrapper[4813]: I0317 09:56:02.433543 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562356-kdgd8" event={"ID":"069b5d80-8eea-46f2-b7e2-1843d75656d3","Type":"ContainerStarted","Data":"3c6beb0903e9cbf86086e59294c2f7965cb0717b4db6da5907078a267282ded1"} Mar 17 09:56:03 crc kubenswrapper[4813]: I0317 09:56:03.445347 4813 generic.go:334] "Generic (PLEG): container finished" podID="069b5d80-8eea-46f2-b7e2-1843d75656d3" containerID="f5b7f59921e77cafb74627590ca560bbc60aef6fbba2a9c48cc4a148af76c0b4" exitCode=0 Mar 17 09:56:03 crc kubenswrapper[4813]: I0317 09:56:03.445442 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562356-kdgd8" event={"ID":"069b5d80-8eea-46f2-b7e2-1843d75656d3","Type":"ContainerDied","Data":"f5b7f59921e77cafb74627590ca560bbc60aef6fbba2a9c48cc4a148af76c0b4"} Mar 17 09:56:04 crc kubenswrapper[4813]: I0317 09:56:04.894989 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562356-kdgd8" Mar 17 09:56:04 crc kubenswrapper[4813]: I0317 09:56:04.945570 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g4bvh\" (UniqueName: \"kubernetes.io/projected/069b5d80-8eea-46f2-b7e2-1843d75656d3-kube-api-access-g4bvh\") pod \"069b5d80-8eea-46f2-b7e2-1843d75656d3\" (UID: \"069b5d80-8eea-46f2-b7e2-1843d75656d3\") " Mar 17 09:56:04 crc kubenswrapper[4813]: I0317 09:56:04.951770 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/069b5d80-8eea-46f2-b7e2-1843d75656d3-kube-api-access-g4bvh" (OuterVolumeSpecName: "kube-api-access-g4bvh") pod "069b5d80-8eea-46f2-b7e2-1843d75656d3" (UID: "069b5d80-8eea-46f2-b7e2-1843d75656d3"). InnerVolumeSpecName "kube-api-access-g4bvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:56:05 crc kubenswrapper[4813]: I0317 09:56:05.047468 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g4bvh\" (UniqueName: \"kubernetes.io/projected/069b5d80-8eea-46f2-b7e2-1843d75656d3-kube-api-access-g4bvh\") on node \"crc\" DevicePath \"\"" Mar 17 09:56:05 crc kubenswrapper[4813]: I0317 09:56:05.474541 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562356-kdgd8" event={"ID":"069b5d80-8eea-46f2-b7e2-1843d75656d3","Type":"ContainerDied","Data":"3c6beb0903e9cbf86086e59294c2f7965cb0717b4db6da5907078a267282ded1"} Mar 17 09:56:05 crc kubenswrapper[4813]: I0317 09:56:05.475050 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3c6beb0903e9cbf86086e59294c2f7965cb0717b4db6da5907078a267282ded1" Mar 17 09:56:05 crc kubenswrapper[4813]: I0317 09:56:05.474868 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562356-kdgd8" Mar 17 09:56:05 crc kubenswrapper[4813]: I0317 09:56:05.973018 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562350-6hx9r"] Mar 17 09:56:05 crc kubenswrapper[4813]: I0317 09:56:05.981422 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562350-6hx9r"] Mar 17 09:56:06 crc kubenswrapper[4813]: I0317 09:56:06.752902 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d125196b-08fb-4b13-9930-daf241996e49" path="/var/lib/kubelet/pods/d125196b-08fb-4b13-9930-daf241996e49/volumes" Mar 17 09:56:12 crc kubenswrapper[4813]: I0317 09:56:12.592139 4813 generic.go:334] "Generic (PLEG): container finished" podID="cd1a58de-c5c8-4b15-a894-47baac35d6e2" containerID="afd527dab194ddc87ce9959e18af104a344ca53b33c34a66310db549bf3bbc76" exitCode=0 Mar 17 09:56:12 crc kubenswrapper[4813]: I0317 09:56:12.592280 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" event={"ID":"cd1a58de-c5c8-4b15-a894-47baac35d6e2","Type":"ContainerDied","Data":"afd527dab194ddc87ce9959e18af104a344ca53b33c34a66310db549bf3bbc76"} Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.057196 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.114264 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.114347 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.254566 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ssh-key-openstack-edpm-ipam\") pod \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.254785 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ceilometer-compute-config-data-1\") pod \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.254826 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-telemetry-combined-ca-bundle\") pod \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.254901 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-inventory\") pod \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.254957 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ceilometer-compute-config-data-0\") pod \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.255795 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lt7s2\" (UniqueName: \"kubernetes.io/projected/cd1a58de-c5c8-4b15-a894-47baac35d6e2-kube-api-access-lt7s2\") pod \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.256413 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ceilometer-compute-config-data-2\") pod \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\" (UID: \"cd1a58de-c5c8-4b15-a894-47baac35d6e2\") " Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.272047 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd1a58de-c5c8-4b15-a894-47baac35d6e2-kube-api-access-lt7s2" (OuterVolumeSpecName: "kube-api-access-lt7s2") pod "cd1a58de-c5c8-4b15-a894-47baac35d6e2" (UID: "cd1a58de-c5c8-4b15-a894-47baac35d6e2"). InnerVolumeSpecName "kube-api-access-lt7s2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.272954 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "cd1a58de-c5c8-4b15-a894-47baac35d6e2" (UID: "cd1a58de-c5c8-4b15-a894-47baac35d6e2"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.291225 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "cd1a58de-c5c8-4b15-a894-47baac35d6e2" (UID: "cd1a58de-c5c8-4b15-a894-47baac35d6e2"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.310371 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-inventory" (OuterVolumeSpecName: "inventory") pod "cd1a58de-c5c8-4b15-a894-47baac35d6e2" (UID: "cd1a58de-c5c8-4b15-a894-47baac35d6e2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.311149 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "cd1a58de-c5c8-4b15-a894-47baac35d6e2" (UID: "cd1a58de-c5c8-4b15-a894-47baac35d6e2"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.312950 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "cd1a58de-c5c8-4b15-a894-47baac35d6e2" (UID: "cd1a58de-c5c8-4b15-a894-47baac35d6e2"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.315791 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "cd1a58de-c5c8-4b15-a894-47baac35d6e2" (UID: "cd1a58de-c5c8-4b15-a894-47baac35d6e2"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.358472 4813 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.358509 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lt7s2\" (UniqueName: \"kubernetes.io/projected/cd1a58de-c5c8-4b15-a894-47baac35d6e2-kube-api-access-lt7s2\") on node \"crc\" DevicePath \"\"" Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.358528 4813 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.358542 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.358554 4813 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.358567 4813 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.358582 4813 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/cd1a58de-c5c8-4b15-a894-47baac35d6e2-inventory\") on node \"crc\" DevicePath \"\"" Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.615931 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" event={"ID":"cd1a58de-c5c8-4b15-a894-47baac35d6e2","Type":"ContainerDied","Data":"64a84a76c5d479fde74df9186af00359d6a78efad02a1554d95353e8b8e2d239"} Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.615977 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="64a84a76c5d479fde74df9186af00359d6a78efad02a1554d95353e8b8e2d239" Mar 17 09:56:14 crc kubenswrapper[4813]: I0317 09:56:14.615990 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-972kp" Mar 17 09:56:20 crc kubenswrapper[4813]: I0317 09:56:20.481825 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-pr6tk"] Mar 17 09:56:20 crc kubenswrapper[4813]: E0317 09:56:20.482856 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="069b5d80-8eea-46f2-b7e2-1843d75656d3" containerName="oc" Mar 17 09:56:20 crc kubenswrapper[4813]: I0317 09:56:20.482876 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="069b5d80-8eea-46f2-b7e2-1843d75656d3" containerName="oc" Mar 17 09:56:20 crc kubenswrapper[4813]: E0317 09:56:20.482916 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd1a58de-c5c8-4b15-a894-47baac35d6e2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 17 09:56:20 crc kubenswrapper[4813]: I0317 09:56:20.482925 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd1a58de-c5c8-4b15-a894-47baac35d6e2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 17 09:56:20 crc kubenswrapper[4813]: I0317 09:56:20.483141 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd1a58de-c5c8-4b15-a894-47baac35d6e2" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Mar 17 09:56:20 crc kubenswrapper[4813]: I0317 09:56:20.483158 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="069b5d80-8eea-46f2-b7e2-1843d75656d3" containerName="oc" Mar 17 09:56:20 crc kubenswrapper[4813]: I0317 09:56:20.487403 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pr6tk" Mar 17 09:56:20 crc kubenswrapper[4813]: I0317 09:56:20.499635 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pr6tk"] Mar 17 09:56:20 crc kubenswrapper[4813]: I0317 09:56:20.510498 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8b71cf5-0aee-4d4e-a861-bb1bb52fe746-utilities\") pod \"community-operators-pr6tk\" (UID: \"f8b71cf5-0aee-4d4e-a861-bb1bb52fe746\") " pod="openshift-marketplace/community-operators-pr6tk" Mar 17 09:56:20 crc kubenswrapper[4813]: I0317 09:56:20.510537 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zth5b\" (UniqueName: \"kubernetes.io/projected/f8b71cf5-0aee-4d4e-a861-bb1bb52fe746-kube-api-access-zth5b\") pod \"community-operators-pr6tk\" (UID: \"f8b71cf5-0aee-4d4e-a861-bb1bb52fe746\") " pod="openshift-marketplace/community-operators-pr6tk" Mar 17 09:56:20 crc kubenswrapper[4813]: I0317 09:56:20.510642 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8b71cf5-0aee-4d4e-a861-bb1bb52fe746-catalog-content\") pod \"community-operators-pr6tk\" (UID: \"f8b71cf5-0aee-4d4e-a861-bb1bb52fe746\") " pod="openshift-marketplace/community-operators-pr6tk" Mar 17 09:56:20 crc kubenswrapper[4813]: I0317 09:56:20.612244 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8b71cf5-0aee-4d4e-a861-bb1bb52fe746-utilities\") pod \"community-operators-pr6tk\" (UID: \"f8b71cf5-0aee-4d4e-a861-bb1bb52fe746\") " pod="openshift-marketplace/community-operators-pr6tk" Mar 17 09:56:20 crc kubenswrapper[4813]: I0317 09:56:20.612300 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zth5b\" (UniqueName: \"kubernetes.io/projected/f8b71cf5-0aee-4d4e-a861-bb1bb52fe746-kube-api-access-zth5b\") pod \"community-operators-pr6tk\" (UID: \"f8b71cf5-0aee-4d4e-a861-bb1bb52fe746\") " pod="openshift-marketplace/community-operators-pr6tk" Mar 17 09:56:20 crc kubenswrapper[4813]: I0317 09:56:20.612557 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8b71cf5-0aee-4d4e-a861-bb1bb52fe746-catalog-content\") pod \"community-operators-pr6tk\" (UID: \"f8b71cf5-0aee-4d4e-a861-bb1bb52fe746\") " pod="openshift-marketplace/community-operators-pr6tk" Mar 17 09:56:20 crc kubenswrapper[4813]: I0317 09:56:20.612935 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8b71cf5-0aee-4d4e-a861-bb1bb52fe746-utilities\") pod \"community-operators-pr6tk\" (UID: \"f8b71cf5-0aee-4d4e-a861-bb1bb52fe746\") " pod="openshift-marketplace/community-operators-pr6tk" Mar 17 09:56:20 crc kubenswrapper[4813]: I0317 09:56:20.613141 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8b71cf5-0aee-4d4e-a861-bb1bb52fe746-catalog-content\") pod \"community-operators-pr6tk\" (UID: \"f8b71cf5-0aee-4d4e-a861-bb1bb52fe746\") " pod="openshift-marketplace/community-operators-pr6tk" Mar 17 09:56:20 crc kubenswrapper[4813]: I0317 09:56:20.637067 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zth5b\" (UniqueName: \"kubernetes.io/projected/f8b71cf5-0aee-4d4e-a861-bb1bb52fe746-kube-api-access-zth5b\") pod \"community-operators-pr6tk\" (UID: \"f8b71cf5-0aee-4d4e-a861-bb1bb52fe746\") " pod="openshift-marketplace/community-operators-pr6tk" Mar 17 09:56:20 crc kubenswrapper[4813]: I0317 09:56:20.831853 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-pr6tk" Mar 17 09:56:21 crc kubenswrapper[4813]: I0317 09:56:21.391243 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pr6tk"] Mar 17 09:56:21 crc kubenswrapper[4813]: I0317 09:56:21.710527 4813 generic.go:334] "Generic (PLEG): container finished" podID="f8b71cf5-0aee-4d4e-a861-bb1bb52fe746" containerID="938db01e8013bf4b8fe94dbead4af25360ec5ee2f913a5cf92ae1bcde06df89b" exitCode=0 Mar 17 09:56:21 crc kubenswrapper[4813]: I0317 09:56:21.710932 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pr6tk" event={"ID":"f8b71cf5-0aee-4d4e-a861-bb1bb52fe746","Type":"ContainerDied","Data":"938db01e8013bf4b8fe94dbead4af25360ec5ee2f913a5cf92ae1bcde06df89b"} Mar 17 09:56:21 crc kubenswrapper[4813]: I0317 09:56:21.710970 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pr6tk" event={"ID":"f8b71cf5-0aee-4d4e-a861-bb1bb52fe746","Type":"ContainerStarted","Data":"7586b69bb9a30fb82aebdf52ed669f587dc713b5b71fb5be21be4b701301d931"} Mar 17 09:56:25 crc kubenswrapper[4813]: I0317 09:56:25.752368 4813 generic.go:334] "Generic (PLEG): container finished" podID="f8b71cf5-0aee-4d4e-a861-bb1bb52fe746" containerID="dbd40bba524d5e45e401f139de55ebe13bad18eea082e5ab8129b3e7ae299751" exitCode=0 Mar 17 09:56:25 crc kubenswrapper[4813]: I0317 09:56:25.752512 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pr6tk" event={"ID":"f8b71cf5-0aee-4d4e-a861-bb1bb52fe746","Type":"ContainerDied","Data":"dbd40bba524d5e45e401f139de55ebe13bad18eea082e5ab8129b3e7ae299751"} Mar 17 09:56:26 crc kubenswrapper[4813]: I0317 09:56:26.766363 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-pr6tk" event={"ID":"f8b71cf5-0aee-4d4e-a861-bb1bb52fe746","Type":"ContainerStarted","Data":"f926358645f763a99744a33ad2a23295820c0842fac5b65bbae3b773dab63ebb"} Mar 17 09:56:26 crc kubenswrapper[4813]: I0317 09:56:26.803946 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-pr6tk" podStartSLOduration=2.29388167 podStartE2EDuration="6.803916289s" podCreationTimestamp="2026-03-17 09:56:20 +0000 UTC" firstStartedPulling="2026-03-17 09:56:21.712890693 +0000 UTC m=+2803.813694232" lastFinishedPulling="2026-03-17 09:56:26.222925312 +0000 UTC m=+2808.323728851" observedRunningTime="2026-03-17 09:56:26.794873698 +0000 UTC m=+2808.895677187" watchObservedRunningTime="2026-03-17 09:56:26.803916289 +0000 UTC m=+2808.904719818" Mar 17 09:56:30 crc kubenswrapper[4813]: I0317 09:56:30.832143 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-pr6tk" Mar 17 09:56:30 crc kubenswrapper[4813]: I0317 09:56:30.832790 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-pr6tk" Mar 17 09:56:30 crc kubenswrapper[4813]: I0317 09:56:30.924074 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-pr6tk" Mar 17 09:56:31 crc kubenswrapper[4813]: I0317 09:56:31.892964 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-pr6tk" Mar 17 09:56:31 crc kubenswrapper[4813]: I0317 09:56:31.990391 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-pr6tk"] Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.032347 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rshd8"] Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.032667 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-rshd8" podUID="a972da79-2f83-434c-8466-d59b21923a84" containerName="registry-server" containerID="cri-o://83013ad1063315714c2b36d68b2f80061a8950bdfa29ed7fd9c4c370263e506f" gracePeriod=2 Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.528799 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rshd8" Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.677809 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a972da79-2f83-434c-8466-d59b21923a84-utilities\") pod \"a972da79-2f83-434c-8466-d59b21923a84\" (UID: \"a972da79-2f83-434c-8466-d59b21923a84\") " Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.677966 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a972da79-2f83-434c-8466-d59b21923a84-catalog-content\") pod \"a972da79-2f83-434c-8466-d59b21923a84\" (UID: \"a972da79-2f83-434c-8466-d59b21923a84\") " Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.678035 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8rzv2\" (UniqueName: \"kubernetes.io/projected/a972da79-2f83-434c-8466-d59b21923a84-kube-api-access-8rzv2\") pod \"a972da79-2f83-434c-8466-d59b21923a84\" (UID: \"a972da79-2f83-434c-8466-d59b21923a84\") " Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.678374 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a972da79-2f83-434c-8466-d59b21923a84-utilities" (OuterVolumeSpecName: "utilities") pod "a972da79-2f83-434c-8466-d59b21923a84" (UID: "a972da79-2f83-434c-8466-d59b21923a84"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.678461 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a972da79-2f83-434c-8466-d59b21923a84-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.690827 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a972da79-2f83-434c-8466-d59b21923a84-kube-api-access-8rzv2" (OuterVolumeSpecName: "kube-api-access-8rzv2") pod "a972da79-2f83-434c-8466-d59b21923a84" (UID: "a972da79-2f83-434c-8466-d59b21923a84"). InnerVolumeSpecName "kube-api-access-8rzv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.729068 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a972da79-2f83-434c-8466-d59b21923a84-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a972da79-2f83-434c-8466-d59b21923a84" (UID: "a972da79-2f83-434c-8466-d59b21923a84"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.780163 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a972da79-2f83-434c-8466-d59b21923a84-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.780201 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8rzv2\" (UniqueName: \"kubernetes.io/projected/a972da79-2f83-434c-8466-d59b21923a84-kube-api-access-8rzv2\") on node \"crc\" DevicePath \"\"" Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.833467 4813 generic.go:334] "Generic (PLEG): container finished" podID="a972da79-2f83-434c-8466-d59b21923a84" containerID="83013ad1063315714c2b36d68b2f80061a8950bdfa29ed7fd9c4c370263e506f" exitCode=0 Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.833552 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rshd8" event={"ID":"a972da79-2f83-434c-8466-d59b21923a84","Type":"ContainerDied","Data":"83013ad1063315714c2b36d68b2f80061a8950bdfa29ed7fd9c4c370263e506f"} Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.833635 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rshd8" event={"ID":"a972da79-2f83-434c-8466-d59b21923a84","Type":"ContainerDied","Data":"dfcc7321fb92c0213f31c83e00833698ba1f889f3dfb35063062181c9673922e"} Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.833654 4813 scope.go:117] "RemoveContainer" containerID="83013ad1063315714c2b36d68b2f80061a8950bdfa29ed7fd9c4c370263e506f" Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.833892 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rshd8" Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.858543 4813 scope.go:117] "RemoveContainer" containerID="c9bd84b9903ae3f96730a1048c03f6d9da637aaf6cab6649578e9e4bc4f99d18" Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.862583 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-rshd8"] Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.873679 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-rshd8"] Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.905387 4813 scope.go:117] "RemoveContainer" containerID="821d50bf0926f340da658e2ab21b2189e3de607da0f6c3df7209cd6222ec2db4" Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.929042 4813 scope.go:117] "RemoveContainer" containerID="83013ad1063315714c2b36d68b2f80061a8950bdfa29ed7fd9c4c370263e506f" Mar 17 09:56:32 crc kubenswrapper[4813]: E0317 09:56:32.929818 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"83013ad1063315714c2b36d68b2f80061a8950bdfa29ed7fd9c4c370263e506f\": container with ID starting with 83013ad1063315714c2b36d68b2f80061a8950bdfa29ed7fd9c4c370263e506f not found: ID does not exist" containerID="83013ad1063315714c2b36d68b2f80061a8950bdfa29ed7fd9c4c370263e506f" Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.929856 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"83013ad1063315714c2b36d68b2f80061a8950bdfa29ed7fd9c4c370263e506f"} err="failed to get container status \"83013ad1063315714c2b36d68b2f80061a8950bdfa29ed7fd9c4c370263e506f\": rpc error: code = NotFound desc = could not find container \"83013ad1063315714c2b36d68b2f80061a8950bdfa29ed7fd9c4c370263e506f\": container with ID starting with 83013ad1063315714c2b36d68b2f80061a8950bdfa29ed7fd9c4c370263e506f not found: ID does not exist" Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.929882 4813 scope.go:117] "RemoveContainer" containerID="c9bd84b9903ae3f96730a1048c03f6d9da637aaf6cab6649578e9e4bc4f99d18" Mar 17 09:56:32 crc kubenswrapper[4813]: E0317 09:56:32.930163 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9bd84b9903ae3f96730a1048c03f6d9da637aaf6cab6649578e9e4bc4f99d18\": container with ID starting with c9bd84b9903ae3f96730a1048c03f6d9da637aaf6cab6649578e9e4bc4f99d18 not found: ID does not exist" containerID="c9bd84b9903ae3f96730a1048c03f6d9da637aaf6cab6649578e9e4bc4f99d18" Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.930188 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9bd84b9903ae3f96730a1048c03f6d9da637aaf6cab6649578e9e4bc4f99d18"} err="failed to get container status \"c9bd84b9903ae3f96730a1048c03f6d9da637aaf6cab6649578e9e4bc4f99d18\": rpc error: code = NotFound desc = could not find container \"c9bd84b9903ae3f96730a1048c03f6d9da637aaf6cab6649578e9e4bc4f99d18\": container with ID starting with c9bd84b9903ae3f96730a1048c03f6d9da637aaf6cab6649578e9e4bc4f99d18 not found: ID does not exist" Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.930201 4813 scope.go:117] "RemoveContainer" containerID="821d50bf0926f340da658e2ab21b2189e3de607da0f6c3df7209cd6222ec2db4" Mar 17 09:56:32 crc kubenswrapper[4813]: E0317 09:56:32.933199 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"821d50bf0926f340da658e2ab21b2189e3de607da0f6c3df7209cd6222ec2db4\": container with ID starting with 821d50bf0926f340da658e2ab21b2189e3de607da0f6c3df7209cd6222ec2db4 not found: ID does not exist" containerID="821d50bf0926f340da658e2ab21b2189e3de607da0f6c3df7209cd6222ec2db4" Mar 17 09:56:32 crc kubenswrapper[4813]: I0317 09:56:32.933230 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"821d50bf0926f340da658e2ab21b2189e3de607da0f6c3df7209cd6222ec2db4"} err="failed to get container status \"821d50bf0926f340da658e2ab21b2189e3de607da0f6c3df7209cd6222ec2db4\": rpc error: code = NotFound desc = could not find container \"821d50bf0926f340da658e2ab21b2189e3de607da0f6c3df7209cd6222ec2db4\": container with ID starting with 821d50bf0926f340da658e2ab21b2189e3de607da0f6c3df7209cd6222ec2db4 not found: ID does not exist" Mar 17 09:56:34 crc kubenswrapper[4813]: I0317 09:56:34.744797 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a972da79-2f83-434c-8466-d59b21923a84" path="/var/lib/kubelet/pods/a972da79-2f83-434c-8466-d59b21923a84/volumes" Mar 17 09:56:44 crc kubenswrapper[4813]: I0317 09:56:44.114462 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:56:44 crc kubenswrapper[4813]: I0317 09:56:44.115419 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:56:44 crc kubenswrapper[4813]: I0317 09:56:44.115515 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:56:44 crc kubenswrapper[4813]: I0317 09:56:44.116882 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"df44ca5fc960279ac359c596c0c46ba2f6c33c03061a11631b7e3d791130963b"} pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 09:56:44 crc kubenswrapper[4813]: I0317 09:56:44.117018 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" containerID="cri-o://df44ca5fc960279ac359c596c0c46ba2f6c33c03061a11631b7e3d791130963b" gracePeriod=600 Mar 17 09:56:44 crc kubenswrapper[4813]: I0317 09:56:44.973741 4813 generic.go:334] "Generic (PLEG): container finished" podID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerID="df44ca5fc960279ac359c596c0c46ba2f6c33c03061a11631b7e3d791130963b" exitCode=0 Mar 17 09:56:44 crc kubenswrapper[4813]: I0317 09:56:44.973798 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerDied","Data":"df44ca5fc960279ac359c596c0c46ba2f6c33c03061a11631b7e3d791130963b"} Mar 17 09:56:44 crc kubenswrapper[4813]: I0317 09:56:44.974178 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerStarted","Data":"9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677"} Mar 17 09:56:44 crc kubenswrapper[4813]: I0317 09:56:44.974213 4813 scope.go:117] "RemoveContainer" containerID="5ae09e910bb4f07323b1b208350159c23425fea564d85a6a8957e0bf1b33e8c4" Mar 17 09:56:49 crc kubenswrapper[4813]: I0317 09:56:49.561197 4813 scope.go:117] "RemoveContainer" containerID="806dd70c01c034a744ce0d93e725c97fa79fc217ed0954e2d050eaf82a87e54e" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.390161 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Mar 17 09:57:13 crc kubenswrapper[4813]: E0317 09:57:13.392072 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a972da79-2f83-434c-8466-d59b21923a84" containerName="registry-server" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.392104 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a972da79-2f83-434c-8466-d59b21923a84" containerName="registry-server" Mar 17 09:57:13 crc kubenswrapper[4813]: E0317 09:57:13.392120 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a972da79-2f83-434c-8466-d59b21923a84" containerName="extract-utilities" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.392128 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a972da79-2f83-434c-8466-d59b21923a84" containerName="extract-utilities" Mar 17 09:57:13 crc kubenswrapper[4813]: E0317 09:57:13.392147 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a972da79-2f83-434c-8466-d59b21923a84" containerName="extract-content" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.392154 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a972da79-2f83-434c-8466-d59b21923a84" containerName="extract-content" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.392387 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a972da79-2f83-434c-8466-d59b21923a84" containerName="registry-server" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.393006 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.395542 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.395723 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.397573 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-4lvrg" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.404474 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.409249 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.481901 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/005f04b0-01f5-4bed-9d2c-2269ab67d27f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.482038 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/005f04b0-01f5-4bed-9d2c-2269ab67d27f-config-data\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.482098 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/005f04b0-01f5-4bed-9d2c-2269ab67d27f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.482124 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/005f04b0-01f5-4bed-9d2c-2269ab67d27f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.482291 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.482440 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwb74\" (UniqueName: \"kubernetes.io/projected/005f04b0-01f5-4bed-9d2c-2269ab67d27f-kube-api-access-fwb74\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.482511 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/005f04b0-01f5-4bed-9d2c-2269ab67d27f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.482563 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/005f04b0-01f5-4bed-9d2c-2269ab67d27f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.482622 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/005f04b0-01f5-4bed-9d2c-2269ab67d27f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.584419 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/005f04b0-01f5-4bed-9d2c-2269ab67d27f-config-data\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.584522 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/005f04b0-01f5-4bed-9d2c-2269ab67d27f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.584571 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/005f04b0-01f5-4bed-9d2c-2269ab67d27f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.584657 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.584688 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwb74\" (UniqueName: \"kubernetes.io/projected/005f04b0-01f5-4bed-9d2c-2269ab67d27f-kube-api-access-fwb74\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.584727 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/005f04b0-01f5-4bed-9d2c-2269ab67d27f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.584764 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/005f04b0-01f5-4bed-9d2c-2269ab67d27f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.584812 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/005f04b0-01f5-4bed-9d2c-2269ab67d27f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.584968 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/005f04b0-01f5-4bed-9d2c-2269ab67d27f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.585458 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/005f04b0-01f5-4bed-9d2c-2269ab67d27f-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.585657 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.586908 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/005f04b0-01f5-4bed-9d2c-2269ab67d27f-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.587436 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/005f04b0-01f5-4bed-9d2c-2269ab67d27f-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.587755 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/005f04b0-01f5-4bed-9d2c-2269ab67d27f-config-data\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.596188 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/005f04b0-01f5-4bed-9d2c-2269ab67d27f-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.596445 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/005f04b0-01f5-4bed-9d2c-2269ab67d27f-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.596565 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/005f04b0-01f5-4bed-9d2c-2269ab67d27f-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.604506 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwb74\" (UniqueName: \"kubernetes.io/projected/005f04b0-01f5-4bed-9d2c-2269ab67d27f-kube-api-access-fwb74\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.639003 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"tempest-tests-tempest\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " pod="openstack/tempest-tests-tempest" Mar 17 09:57:13 crc kubenswrapper[4813]: I0317 09:57:13.724776 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 17 09:57:14 crc kubenswrapper[4813]: I0317 09:57:14.170636 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Mar 17 09:57:14 crc kubenswrapper[4813]: I0317 09:57:14.171684 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 09:57:14 crc kubenswrapper[4813]: I0317 09:57:14.695791 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"005f04b0-01f5-4bed-9d2c-2269ab67d27f","Type":"ContainerStarted","Data":"41d3487d7af3eae5152a5532d1c7b33b8d7a4e1b8abda035bb1321054aac8d7a"} Mar 17 09:57:42 crc kubenswrapper[4813]: E0317 09:57:42.221484 4813 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Mar 17 09:57:42 crc kubenswrapper[4813]: E0317 09:57:42.222260 4813 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fwb74,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(005f04b0-01f5-4bed-9d2c-2269ab67d27f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Mar 17 09:57:42 crc kubenswrapper[4813]: E0317 09:57:42.223524 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="005f04b0-01f5-4bed-9d2c-2269ab67d27f" Mar 17 09:57:42 crc kubenswrapper[4813]: E0317 09:57:42.982702 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="005f04b0-01f5-4bed-9d2c-2269ab67d27f" Mar 17 09:57:56 crc kubenswrapper[4813]: I0317 09:57:56.526125 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Mar 17 09:57:59 crc kubenswrapper[4813]: I0317 09:57:59.523448 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"005f04b0-01f5-4bed-9d2c-2269ab67d27f","Type":"ContainerStarted","Data":"a645cf679eaec3380a711eb52ff7a30b787011ec69a129842da78db374bf1842"} Mar 17 09:57:59 crc kubenswrapper[4813]: I0317 09:57:59.558627 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=5.208026278 podStartE2EDuration="47.558578732s" podCreationTimestamp="2026-03-17 09:57:12 +0000 UTC" firstStartedPulling="2026-03-17 09:57:14.171426607 +0000 UTC m=+2856.272230116" lastFinishedPulling="2026-03-17 09:57:56.521979041 +0000 UTC m=+2898.622782570" observedRunningTime="2026-03-17 09:57:59.547423484 +0000 UTC m=+2901.648227023" watchObservedRunningTime="2026-03-17 09:57:59.558578732 +0000 UTC m=+2901.659382241" Mar 17 09:58:00 crc kubenswrapper[4813]: I0317 09:58:00.152422 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562358-kdqtd"] Mar 17 09:58:00 crc kubenswrapper[4813]: I0317 09:58:00.154648 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562358-kdqtd" Mar 17 09:58:00 crc kubenswrapper[4813]: I0317 09:58:00.156827 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 09:58:00 crc kubenswrapper[4813]: I0317 09:58:00.157391 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 09:58:00 crc kubenswrapper[4813]: I0317 09:58:00.161014 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 09:58:00 crc kubenswrapper[4813]: I0317 09:58:00.166691 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562358-kdqtd"] Mar 17 09:58:00 crc kubenswrapper[4813]: I0317 09:58:00.288695 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr58x\" (UniqueName: \"kubernetes.io/projected/16fe11ef-273a-403d-8b23-80c3955a869f-kube-api-access-wr58x\") pod \"auto-csr-approver-29562358-kdqtd\" (UID: \"16fe11ef-273a-403d-8b23-80c3955a869f\") " pod="openshift-infra/auto-csr-approver-29562358-kdqtd" Mar 17 09:58:00 crc kubenswrapper[4813]: I0317 09:58:00.394230 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr58x\" (UniqueName: \"kubernetes.io/projected/16fe11ef-273a-403d-8b23-80c3955a869f-kube-api-access-wr58x\") pod \"auto-csr-approver-29562358-kdqtd\" (UID: \"16fe11ef-273a-403d-8b23-80c3955a869f\") " pod="openshift-infra/auto-csr-approver-29562358-kdqtd" Mar 17 09:58:00 crc kubenswrapper[4813]: I0317 09:58:00.427457 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr58x\" (UniqueName: \"kubernetes.io/projected/16fe11ef-273a-403d-8b23-80c3955a869f-kube-api-access-wr58x\") pod \"auto-csr-approver-29562358-kdqtd\" (UID: \"16fe11ef-273a-403d-8b23-80c3955a869f\") " pod="openshift-infra/auto-csr-approver-29562358-kdqtd" Mar 17 09:58:00 crc kubenswrapper[4813]: I0317 09:58:00.473485 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562358-kdqtd" Mar 17 09:58:00 crc kubenswrapper[4813]: I0317 09:58:00.972190 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562358-kdqtd"] Mar 17 09:58:00 crc kubenswrapper[4813]: W0317 09:58:00.972787 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod16fe11ef_273a_403d_8b23_80c3955a869f.slice/crio-4a830d96695a2a3a566456e3f904d47051c12ff4f436477ada5d3add60ee2533 WatchSource:0}: Error finding container 4a830d96695a2a3a566456e3f904d47051c12ff4f436477ada5d3add60ee2533: Status 404 returned error can't find the container with id 4a830d96695a2a3a566456e3f904d47051c12ff4f436477ada5d3add60ee2533 Mar 17 09:58:01 crc kubenswrapper[4813]: I0317 09:58:01.558986 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562358-kdqtd" event={"ID":"16fe11ef-273a-403d-8b23-80c3955a869f","Type":"ContainerStarted","Data":"4a830d96695a2a3a566456e3f904d47051c12ff4f436477ada5d3add60ee2533"} Mar 17 09:58:03 crc kubenswrapper[4813]: I0317 09:58:03.580377 4813 generic.go:334] "Generic (PLEG): container finished" podID="16fe11ef-273a-403d-8b23-80c3955a869f" containerID="eca8ccfa622bf9a226d07a404218fad4fe269b2b795671d0799ed86a629334b5" exitCode=0 Mar 17 09:58:03 crc kubenswrapper[4813]: I0317 09:58:03.580835 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562358-kdqtd" event={"ID":"16fe11ef-273a-403d-8b23-80c3955a869f","Type":"ContainerDied","Data":"eca8ccfa622bf9a226d07a404218fad4fe269b2b795671d0799ed86a629334b5"} Mar 17 09:58:05 crc kubenswrapper[4813]: I0317 09:58:05.004425 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562358-kdqtd" Mar 17 09:58:05 crc kubenswrapper[4813]: I0317 09:58:05.091930 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wr58x\" (UniqueName: \"kubernetes.io/projected/16fe11ef-273a-403d-8b23-80c3955a869f-kube-api-access-wr58x\") pod \"16fe11ef-273a-403d-8b23-80c3955a869f\" (UID: \"16fe11ef-273a-403d-8b23-80c3955a869f\") " Mar 17 09:58:05 crc kubenswrapper[4813]: I0317 09:58:05.107460 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/16fe11ef-273a-403d-8b23-80c3955a869f-kube-api-access-wr58x" (OuterVolumeSpecName: "kube-api-access-wr58x") pod "16fe11ef-273a-403d-8b23-80c3955a869f" (UID: "16fe11ef-273a-403d-8b23-80c3955a869f"). InnerVolumeSpecName "kube-api-access-wr58x". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:58:05 crc kubenswrapper[4813]: I0317 09:58:05.194887 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wr58x\" (UniqueName: \"kubernetes.io/projected/16fe11ef-273a-403d-8b23-80c3955a869f-kube-api-access-wr58x\") on node \"crc\" DevicePath \"\"" Mar 17 09:58:05 crc kubenswrapper[4813]: I0317 09:58:05.612674 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562358-kdqtd" event={"ID":"16fe11ef-273a-403d-8b23-80c3955a869f","Type":"ContainerDied","Data":"4a830d96695a2a3a566456e3f904d47051c12ff4f436477ada5d3add60ee2533"} Mar 17 09:58:05 crc kubenswrapper[4813]: I0317 09:58:05.612913 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4a830d96695a2a3a566456e3f904d47051c12ff4f436477ada5d3add60ee2533" Mar 17 09:58:05 crc kubenswrapper[4813]: I0317 09:58:05.612770 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562358-kdqtd" Mar 17 09:58:06 crc kubenswrapper[4813]: I0317 09:58:06.067799 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562352-n4nx6"] Mar 17 09:58:06 crc kubenswrapper[4813]: I0317 09:58:06.077474 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562352-n4nx6"] Mar 17 09:58:06 crc kubenswrapper[4813]: I0317 09:58:06.741075 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0cb5aa99-7a25-494a-b945-925a88a3f5d7" path="/var/lib/kubelet/pods/0cb5aa99-7a25-494a-b945-925a88a3f5d7/volumes" Mar 17 09:58:44 crc kubenswrapper[4813]: I0317 09:58:44.114578 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:58:44 crc kubenswrapper[4813]: I0317 09:58:44.115350 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:58:49 crc kubenswrapper[4813]: I0317 09:58:49.720781 4813 scope.go:117] "RemoveContainer" containerID="8c153ea76086c7849bf876e64e764357f065dace947978f32afb995852f805e0" Mar 17 09:59:14 crc kubenswrapper[4813]: I0317 09:59:14.114048 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:59:14 crc kubenswrapper[4813]: I0317 09:59:14.114642 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:59:30 crc kubenswrapper[4813]: I0317 09:59:30.118272 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-j6ppb"] Mar 17 09:59:30 crc kubenswrapper[4813]: E0317 09:59:30.119382 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="16fe11ef-273a-403d-8b23-80c3955a869f" containerName="oc" Mar 17 09:59:30 crc kubenswrapper[4813]: I0317 09:59:30.119400 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="16fe11ef-273a-403d-8b23-80c3955a869f" containerName="oc" Mar 17 09:59:30 crc kubenswrapper[4813]: I0317 09:59:30.119733 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="16fe11ef-273a-403d-8b23-80c3955a869f" containerName="oc" Mar 17 09:59:30 crc kubenswrapper[4813]: I0317 09:59:30.121286 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j6ppb" Mar 17 09:59:30 crc kubenswrapper[4813]: I0317 09:59:30.154172 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j6ppb"] Mar 17 09:59:30 crc kubenswrapper[4813]: I0317 09:59:30.243732 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-djwkn\" (UniqueName: \"kubernetes.io/projected/fe4a24a4-5e9c-4955-8612-138ef4b76249-kube-api-access-djwkn\") pod \"redhat-marketplace-j6ppb\" (UID: \"fe4a24a4-5e9c-4955-8612-138ef4b76249\") " pod="openshift-marketplace/redhat-marketplace-j6ppb" Mar 17 09:59:30 crc kubenswrapper[4813]: I0317 09:59:30.243857 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe4a24a4-5e9c-4955-8612-138ef4b76249-catalog-content\") pod \"redhat-marketplace-j6ppb\" (UID: \"fe4a24a4-5e9c-4955-8612-138ef4b76249\") " pod="openshift-marketplace/redhat-marketplace-j6ppb" Mar 17 09:59:30 crc kubenswrapper[4813]: I0317 09:59:30.244330 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe4a24a4-5e9c-4955-8612-138ef4b76249-utilities\") pod \"redhat-marketplace-j6ppb\" (UID: \"fe4a24a4-5e9c-4955-8612-138ef4b76249\") " pod="openshift-marketplace/redhat-marketplace-j6ppb" Mar 17 09:59:30 crc kubenswrapper[4813]: I0317 09:59:30.347688 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-djwkn\" (UniqueName: \"kubernetes.io/projected/fe4a24a4-5e9c-4955-8612-138ef4b76249-kube-api-access-djwkn\") pod \"redhat-marketplace-j6ppb\" (UID: \"fe4a24a4-5e9c-4955-8612-138ef4b76249\") " pod="openshift-marketplace/redhat-marketplace-j6ppb" Mar 17 09:59:30 crc kubenswrapper[4813]: I0317 09:59:30.348742 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe4a24a4-5e9c-4955-8612-138ef4b76249-catalog-content\") pod \"redhat-marketplace-j6ppb\" (UID: \"fe4a24a4-5e9c-4955-8612-138ef4b76249\") " pod="openshift-marketplace/redhat-marketplace-j6ppb" Mar 17 09:59:30 crc kubenswrapper[4813]: I0317 09:59:30.348817 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe4a24a4-5e9c-4955-8612-138ef4b76249-catalog-content\") pod \"redhat-marketplace-j6ppb\" (UID: \"fe4a24a4-5e9c-4955-8612-138ef4b76249\") " pod="openshift-marketplace/redhat-marketplace-j6ppb" Mar 17 09:59:30 crc kubenswrapper[4813]: I0317 09:59:30.349264 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe4a24a4-5e9c-4955-8612-138ef4b76249-utilities\") pod \"redhat-marketplace-j6ppb\" (UID: \"fe4a24a4-5e9c-4955-8612-138ef4b76249\") " pod="openshift-marketplace/redhat-marketplace-j6ppb" Mar 17 09:59:30 crc kubenswrapper[4813]: I0317 09:59:30.349782 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe4a24a4-5e9c-4955-8612-138ef4b76249-utilities\") pod \"redhat-marketplace-j6ppb\" (UID: \"fe4a24a4-5e9c-4955-8612-138ef4b76249\") " pod="openshift-marketplace/redhat-marketplace-j6ppb" Mar 17 09:59:30 crc kubenswrapper[4813]: I0317 09:59:30.366487 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-djwkn\" (UniqueName: \"kubernetes.io/projected/fe4a24a4-5e9c-4955-8612-138ef4b76249-kube-api-access-djwkn\") pod \"redhat-marketplace-j6ppb\" (UID: \"fe4a24a4-5e9c-4955-8612-138ef4b76249\") " pod="openshift-marketplace/redhat-marketplace-j6ppb" Mar 17 09:59:30 crc kubenswrapper[4813]: I0317 09:59:30.467748 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j6ppb" Mar 17 09:59:30 crc kubenswrapper[4813]: I0317 09:59:30.966041 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-j6ppb"] Mar 17 09:59:31 crc kubenswrapper[4813]: I0317 09:59:31.516211 4813 generic.go:334] "Generic (PLEG): container finished" podID="fe4a24a4-5e9c-4955-8612-138ef4b76249" containerID="e58c58d58b257fcd57166bf2adbbdbd11d2dc97c64bcb8f344309bd18cf4c082" exitCode=0 Mar 17 09:59:31 crc kubenswrapper[4813]: I0317 09:59:31.516341 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j6ppb" event={"ID":"fe4a24a4-5e9c-4955-8612-138ef4b76249","Type":"ContainerDied","Data":"e58c58d58b257fcd57166bf2adbbdbd11d2dc97c64bcb8f344309bd18cf4c082"} Mar 17 09:59:31 crc kubenswrapper[4813]: I0317 09:59:31.516530 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j6ppb" event={"ID":"fe4a24a4-5e9c-4955-8612-138ef4b76249","Type":"ContainerStarted","Data":"33afe391688f2bc7551de73309669b63e836a2f8fa4c6f70f79ada85bab5c460"} Mar 17 09:59:32 crc kubenswrapper[4813]: I0317 09:59:32.528495 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j6ppb" event={"ID":"fe4a24a4-5e9c-4955-8612-138ef4b76249","Type":"ContainerStarted","Data":"ce1924d692202b2fbe224f9d0c34940a33ece8b675a0f35ae17c31093bd02f11"} Mar 17 09:59:33 crc kubenswrapper[4813]: I0317 09:59:33.545065 4813 generic.go:334] "Generic (PLEG): container finished" podID="fe4a24a4-5e9c-4955-8612-138ef4b76249" containerID="ce1924d692202b2fbe224f9d0c34940a33ece8b675a0f35ae17c31093bd02f11" exitCode=0 Mar 17 09:59:33 crc kubenswrapper[4813]: I0317 09:59:33.545142 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j6ppb" event={"ID":"fe4a24a4-5e9c-4955-8612-138ef4b76249","Type":"ContainerDied","Data":"ce1924d692202b2fbe224f9d0c34940a33ece8b675a0f35ae17c31093bd02f11"} Mar 17 09:59:34 crc kubenswrapper[4813]: I0317 09:59:34.561748 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j6ppb" event={"ID":"fe4a24a4-5e9c-4955-8612-138ef4b76249","Type":"ContainerStarted","Data":"5c0e1faa090f5996fff32fd36e936ab48758879465699b08c5948e53f6b0f5de"} Mar 17 09:59:34 crc kubenswrapper[4813]: I0317 09:59:34.601396 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-j6ppb" podStartSLOduration=1.93558242 podStartE2EDuration="4.601369397s" podCreationTimestamp="2026-03-17 09:59:30 +0000 UTC" firstStartedPulling="2026-03-17 09:59:31.519381578 +0000 UTC m=+2993.620185087" lastFinishedPulling="2026-03-17 09:59:34.185168515 +0000 UTC m=+2996.285972064" observedRunningTime="2026-03-17 09:59:34.594417852 +0000 UTC m=+2996.695221381" watchObservedRunningTime="2026-03-17 09:59:34.601369397 +0000 UTC m=+2996.702172906" Mar 17 09:59:40 crc kubenswrapper[4813]: I0317 09:59:40.468043 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-j6ppb" Mar 17 09:59:40 crc kubenswrapper[4813]: I0317 09:59:40.468557 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-j6ppb" Mar 17 09:59:40 crc kubenswrapper[4813]: I0317 09:59:40.536171 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-j6ppb" Mar 17 09:59:40 crc kubenswrapper[4813]: I0317 09:59:40.663510 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-j6ppb" Mar 17 09:59:40 crc kubenswrapper[4813]: I0317 09:59:40.767452 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j6ppb"] Mar 17 09:59:42 crc kubenswrapper[4813]: I0317 09:59:42.647749 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-j6ppb" podUID="fe4a24a4-5e9c-4955-8612-138ef4b76249" containerName="registry-server" containerID="cri-o://5c0e1faa090f5996fff32fd36e936ab48758879465699b08c5948e53f6b0f5de" gracePeriod=2 Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.247396 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j6ppb" Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.336222 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-djwkn\" (UniqueName: \"kubernetes.io/projected/fe4a24a4-5e9c-4955-8612-138ef4b76249-kube-api-access-djwkn\") pod \"fe4a24a4-5e9c-4955-8612-138ef4b76249\" (UID: \"fe4a24a4-5e9c-4955-8612-138ef4b76249\") " Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.343849 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fe4a24a4-5e9c-4955-8612-138ef4b76249-kube-api-access-djwkn" (OuterVolumeSpecName: "kube-api-access-djwkn") pod "fe4a24a4-5e9c-4955-8612-138ef4b76249" (UID: "fe4a24a4-5e9c-4955-8612-138ef4b76249"). InnerVolumeSpecName "kube-api-access-djwkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.438247 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe4a24a4-5e9c-4955-8612-138ef4b76249-catalog-content\") pod \"fe4a24a4-5e9c-4955-8612-138ef4b76249\" (UID: \"fe4a24a4-5e9c-4955-8612-138ef4b76249\") " Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.438338 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe4a24a4-5e9c-4955-8612-138ef4b76249-utilities\") pod \"fe4a24a4-5e9c-4955-8612-138ef4b76249\" (UID: \"fe4a24a4-5e9c-4955-8612-138ef4b76249\") " Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.439249 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-djwkn\" (UniqueName: \"kubernetes.io/projected/fe4a24a4-5e9c-4955-8612-138ef4b76249-kube-api-access-djwkn\") on node \"crc\" DevicePath \"\"" Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.440259 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe4a24a4-5e9c-4955-8612-138ef4b76249-utilities" (OuterVolumeSpecName: "utilities") pod "fe4a24a4-5e9c-4955-8612-138ef4b76249" (UID: "fe4a24a4-5e9c-4955-8612-138ef4b76249"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.464818 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fe4a24a4-5e9c-4955-8612-138ef4b76249-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "fe4a24a4-5e9c-4955-8612-138ef4b76249" (UID: "fe4a24a4-5e9c-4955-8612-138ef4b76249"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.540765 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/fe4a24a4-5e9c-4955-8612-138ef4b76249-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.540798 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/fe4a24a4-5e9c-4955-8612-138ef4b76249-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.661935 4813 generic.go:334] "Generic (PLEG): container finished" podID="fe4a24a4-5e9c-4955-8612-138ef4b76249" containerID="5c0e1faa090f5996fff32fd36e936ab48758879465699b08c5948e53f6b0f5de" exitCode=0 Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.661993 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j6ppb" event={"ID":"fe4a24a4-5e9c-4955-8612-138ef4b76249","Type":"ContainerDied","Data":"5c0e1faa090f5996fff32fd36e936ab48758879465699b08c5948e53f6b0f5de"} Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.662011 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-j6ppb" Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.662040 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-j6ppb" event={"ID":"fe4a24a4-5e9c-4955-8612-138ef4b76249","Type":"ContainerDied","Data":"33afe391688f2bc7551de73309669b63e836a2f8fa4c6f70f79ada85bab5c460"} Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.662067 4813 scope.go:117] "RemoveContainer" containerID="5c0e1faa090f5996fff32fd36e936ab48758879465699b08c5948e53f6b0f5de" Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.703909 4813 scope.go:117] "RemoveContainer" containerID="ce1924d692202b2fbe224f9d0c34940a33ece8b675a0f35ae17c31093bd02f11" Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.709893 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-j6ppb"] Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.720658 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-j6ppb"] Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.736635 4813 scope.go:117] "RemoveContainer" containerID="e58c58d58b257fcd57166bf2adbbdbd11d2dc97c64bcb8f344309bd18cf4c082" Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.801524 4813 scope.go:117] "RemoveContainer" containerID="5c0e1faa090f5996fff32fd36e936ab48758879465699b08c5948e53f6b0f5de" Mar 17 09:59:43 crc kubenswrapper[4813]: E0317 09:59:43.802122 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5c0e1faa090f5996fff32fd36e936ab48758879465699b08c5948e53f6b0f5de\": container with ID starting with 5c0e1faa090f5996fff32fd36e936ab48758879465699b08c5948e53f6b0f5de not found: ID does not exist" containerID="5c0e1faa090f5996fff32fd36e936ab48758879465699b08c5948e53f6b0f5de" Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.802258 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5c0e1faa090f5996fff32fd36e936ab48758879465699b08c5948e53f6b0f5de"} err="failed to get container status \"5c0e1faa090f5996fff32fd36e936ab48758879465699b08c5948e53f6b0f5de\": rpc error: code = NotFound desc = could not find container \"5c0e1faa090f5996fff32fd36e936ab48758879465699b08c5948e53f6b0f5de\": container with ID starting with 5c0e1faa090f5996fff32fd36e936ab48758879465699b08c5948e53f6b0f5de not found: ID does not exist" Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.802291 4813 scope.go:117] "RemoveContainer" containerID="ce1924d692202b2fbe224f9d0c34940a33ece8b675a0f35ae17c31093bd02f11" Mar 17 09:59:43 crc kubenswrapper[4813]: E0317 09:59:43.802747 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ce1924d692202b2fbe224f9d0c34940a33ece8b675a0f35ae17c31093bd02f11\": container with ID starting with ce1924d692202b2fbe224f9d0c34940a33ece8b675a0f35ae17c31093bd02f11 not found: ID does not exist" containerID="ce1924d692202b2fbe224f9d0c34940a33ece8b675a0f35ae17c31093bd02f11" Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.802822 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ce1924d692202b2fbe224f9d0c34940a33ece8b675a0f35ae17c31093bd02f11"} err="failed to get container status \"ce1924d692202b2fbe224f9d0c34940a33ece8b675a0f35ae17c31093bd02f11\": rpc error: code = NotFound desc = could not find container \"ce1924d692202b2fbe224f9d0c34940a33ece8b675a0f35ae17c31093bd02f11\": container with ID starting with ce1924d692202b2fbe224f9d0c34940a33ece8b675a0f35ae17c31093bd02f11 not found: ID does not exist" Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.802879 4813 scope.go:117] "RemoveContainer" containerID="e58c58d58b257fcd57166bf2adbbdbd11d2dc97c64bcb8f344309bd18cf4c082" Mar 17 09:59:43 crc kubenswrapper[4813]: E0317 09:59:43.803482 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e58c58d58b257fcd57166bf2adbbdbd11d2dc97c64bcb8f344309bd18cf4c082\": container with ID starting with e58c58d58b257fcd57166bf2adbbdbd11d2dc97c64bcb8f344309bd18cf4c082 not found: ID does not exist" containerID="e58c58d58b257fcd57166bf2adbbdbd11d2dc97c64bcb8f344309bd18cf4c082" Mar 17 09:59:43 crc kubenswrapper[4813]: I0317 09:59:43.803537 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e58c58d58b257fcd57166bf2adbbdbd11d2dc97c64bcb8f344309bd18cf4c082"} err="failed to get container status \"e58c58d58b257fcd57166bf2adbbdbd11d2dc97c64bcb8f344309bd18cf4c082\": rpc error: code = NotFound desc = could not find container \"e58c58d58b257fcd57166bf2adbbdbd11d2dc97c64bcb8f344309bd18cf4c082\": container with ID starting with e58c58d58b257fcd57166bf2adbbdbd11d2dc97c64bcb8f344309bd18cf4c082 not found: ID does not exist" Mar 17 09:59:44 crc kubenswrapper[4813]: I0317 09:59:44.113807 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 09:59:44 crc kubenswrapper[4813]: I0317 09:59:44.114157 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 09:59:44 crc kubenswrapper[4813]: I0317 09:59:44.114219 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 09:59:44 crc kubenswrapper[4813]: I0317 09:59:44.115123 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677"} pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 09:59:44 crc kubenswrapper[4813]: I0317 09:59:44.115200 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" containerID="cri-o://9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" gracePeriod=600 Mar 17 09:59:44 crc kubenswrapper[4813]: E0317 09:59:44.236957 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:59:44 crc kubenswrapper[4813]: I0317 09:59:44.683630 4813 generic.go:334] "Generic (PLEG): container finished" podID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" exitCode=0 Mar 17 09:59:44 crc kubenswrapper[4813]: I0317 09:59:44.683709 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerDied","Data":"9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677"} Mar 17 09:59:44 crc kubenswrapper[4813]: I0317 09:59:44.683794 4813 scope.go:117] "RemoveContainer" containerID="df44ca5fc960279ac359c596c0c46ba2f6c33c03061a11631b7e3d791130963b" Mar 17 09:59:44 crc kubenswrapper[4813]: I0317 09:59:44.685961 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 09:59:44 crc kubenswrapper[4813]: E0317 09:59:44.686224 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 09:59:44 crc kubenswrapper[4813]: I0317 09:59:44.753573 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fe4a24a4-5e9c-4955-8612-138ef4b76249" path="/var/lib/kubelet/pods/fe4a24a4-5e9c-4955-8612-138ef4b76249/volumes" Mar 17 09:59:56 crc kubenswrapper[4813]: I0317 09:59:56.731550 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 09:59:56 crc kubenswrapper[4813]: E0317 09:59:56.733560 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.197248 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562360-k9p9v"] Mar 17 10:00:00 crc kubenswrapper[4813]: E0317 10:00:00.198066 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe4a24a4-5e9c-4955-8612-138ef4b76249" containerName="registry-server" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.198084 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe4a24a4-5e9c-4955-8612-138ef4b76249" containerName="registry-server" Mar 17 10:00:00 crc kubenswrapper[4813]: E0317 10:00:00.198130 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe4a24a4-5e9c-4955-8612-138ef4b76249" containerName="extract-content" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.198137 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe4a24a4-5e9c-4955-8612-138ef4b76249" containerName="extract-content" Mar 17 10:00:00 crc kubenswrapper[4813]: E0317 10:00:00.198151 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fe4a24a4-5e9c-4955-8612-138ef4b76249" containerName="extract-utilities" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.198159 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="fe4a24a4-5e9c-4955-8612-138ef4b76249" containerName="extract-utilities" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.198381 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="fe4a24a4-5e9c-4955-8612-138ef4b76249" containerName="registry-server" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.199167 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562360-k9p9v" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.202086 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.204397 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.206618 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.212074 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562360-zm5lr"] Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.213628 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562360-zm5lr" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.217143 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.218417 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.230658 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562360-k9p9v"] Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.286836 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562360-zm5lr"] Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.374442 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkmq7\" (UniqueName: \"kubernetes.io/projected/0864972d-362c-4157-a956-5b99452985ab-kube-api-access-tkmq7\") pod \"auto-csr-approver-29562360-k9p9v\" (UID: \"0864972d-362c-4157-a956-5b99452985ab\") " pod="openshift-infra/auto-csr-approver-29562360-k9p9v" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.374668 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fks78\" (UniqueName: \"kubernetes.io/projected/915071d1-1575-4b4c-b2fa-1c62db86bce1-kube-api-access-fks78\") pod \"collect-profiles-29562360-zm5lr\" (UID: \"915071d1-1575-4b4c-b2fa-1c62db86bce1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562360-zm5lr" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.374963 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/915071d1-1575-4b4c-b2fa-1c62db86bce1-config-volume\") pod \"collect-profiles-29562360-zm5lr\" (UID: \"915071d1-1575-4b4c-b2fa-1c62db86bce1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562360-zm5lr" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.375054 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/915071d1-1575-4b4c-b2fa-1c62db86bce1-secret-volume\") pod \"collect-profiles-29562360-zm5lr\" (UID: \"915071d1-1575-4b4c-b2fa-1c62db86bce1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562360-zm5lr" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.477536 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fks78\" (UniqueName: \"kubernetes.io/projected/915071d1-1575-4b4c-b2fa-1c62db86bce1-kube-api-access-fks78\") pod \"collect-profiles-29562360-zm5lr\" (UID: \"915071d1-1575-4b4c-b2fa-1c62db86bce1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562360-zm5lr" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.477719 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/915071d1-1575-4b4c-b2fa-1c62db86bce1-config-volume\") pod \"collect-profiles-29562360-zm5lr\" (UID: \"915071d1-1575-4b4c-b2fa-1c62db86bce1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562360-zm5lr" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.477773 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/915071d1-1575-4b4c-b2fa-1c62db86bce1-secret-volume\") pod \"collect-profiles-29562360-zm5lr\" (UID: \"915071d1-1575-4b4c-b2fa-1c62db86bce1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562360-zm5lr" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.477920 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkmq7\" (UniqueName: \"kubernetes.io/projected/0864972d-362c-4157-a956-5b99452985ab-kube-api-access-tkmq7\") pod \"auto-csr-approver-29562360-k9p9v\" (UID: \"0864972d-362c-4157-a956-5b99452985ab\") " pod="openshift-infra/auto-csr-approver-29562360-k9p9v" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.479814 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/915071d1-1575-4b4c-b2fa-1c62db86bce1-config-volume\") pod \"collect-profiles-29562360-zm5lr\" (UID: \"915071d1-1575-4b4c-b2fa-1c62db86bce1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562360-zm5lr" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.486901 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/915071d1-1575-4b4c-b2fa-1c62db86bce1-secret-volume\") pod \"collect-profiles-29562360-zm5lr\" (UID: \"915071d1-1575-4b4c-b2fa-1c62db86bce1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562360-zm5lr" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.496214 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fks78\" (UniqueName: \"kubernetes.io/projected/915071d1-1575-4b4c-b2fa-1c62db86bce1-kube-api-access-fks78\") pod \"collect-profiles-29562360-zm5lr\" (UID: \"915071d1-1575-4b4c-b2fa-1c62db86bce1\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562360-zm5lr" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.497148 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkmq7\" (UniqueName: \"kubernetes.io/projected/0864972d-362c-4157-a956-5b99452985ab-kube-api-access-tkmq7\") pod \"auto-csr-approver-29562360-k9p9v\" (UID: \"0864972d-362c-4157-a956-5b99452985ab\") " pod="openshift-infra/auto-csr-approver-29562360-k9p9v" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.519114 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562360-k9p9v" Mar 17 10:00:00 crc kubenswrapper[4813]: I0317 10:00:00.537903 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562360-zm5lr" Mar 17 10:00:01 crc kubenswrapper[4813]: I0317 10:00:01.042873 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562360-zm5lr"] Mar 17 10:00:01 crc kubenswrapper[4813]: I0317 10:00:01.109305 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562360-k9p9v"] Mar 17 10:00:01 crc kubenswrapper[4813]: W0317 10:00:01.126732 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0864972d_362c_4157_a956_5b99452985ab.slice/crio-2150fe11a9cae4d84d4e36bf76b15b93bcd77c2e2bd139988ce28ecd81259abe WatchSource:0}: Error finding container 2150fe11a9cae4d84d4e36bf76b15b93bcd77c2e2bd139988ce28ecd81259abe: Status 404 returned error can't find the container with id 2150fe11a9cae4d84d4e36bf76b15b93bcd77c2e2bd139988ce28ecd81259abe Mar 17 10:00:01 crc kubenswrapper[4813]: I0317 10:00:01.899980 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562360-k9p9v" event={"ID":"0864972d-362c-4157-a956-5b99452985ab","Type":"ContainerStarted","Data":"2150fe11a9cae4d84d4e36bf76b15b93bcd77c2e2bd139988ce28ecd81259abe"} Mar 17 10:00:01 crc kubenswrapper[4813]: I0317 10:00:01.901686 4813 generic.go:334] "Generic (PLEG): container finished" podID="915071d1-1575-4b4c-b2fa-1c62db86bce1" containerID="6373a7557f9183a76077ca6e571bf4dddd65bd359013ec3a285b27cd8785b3da" exitCode=0 Mar 17 10:00:01 crc kubenswrapper[4813]: I0317 10:00:01.901716 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562360-zm5lr" event={"ID":"915071d1-1575-4b4c-b2fa-1c62db86bce1","Type":"ContainerDied","Data":"6373a7557f9183a76077ca6e571bf4dddd65bd359013ec3a285b27cd8785b3da"} Mar 17 10:00:01 crc kubenswrapper[4813]: I0317 10:00:01.901759 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562360-zm5lr" event={"ID":"915071d1-1575-4b4c-b2fa-1c62db86bce1","Type":"ContainerStarted","Data":"302f0edcc364ff2e7c5589c43be6fe5f48197111b766549f0aa0abc7712ceb78"} Mar 17 10:00:03 crc kubenswrapper[4813]: I0317 10:00:03.435574 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562360-zm5lr" Mar 17 10:00:03 crc kubenswrapper[4813]: I0317 10:00:03.548079 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/915071d1-1575-4b4c-b2fa-1c62db86bce1-secret-volume\") pod \"915071d1-1575-4b4c-b2fa-1c62db86bce1\" (UID: \"915071d1-1575-4b4c-b2fa-1c62db86bce1\") " Mar 17 10:00:03 crc kubenswrapper[4813]: I0317 10:00:03.548326 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fks78\" (UniqueName: \"kubernetes.io/projected/915071d1-1575-4b4c-b2fa-1c62db86bce1-kube-api-access-fks78\") pod \"915071d1-1575-4b4c-b2fa-1c62db86bce1\" (UID: \"915071d1-1575-4b4c-b2fa-1c62db86bce1\") " Mar 17 10:00:03 crc kubenswrapper[4813]: I0317 10:00:03.548376 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/915071d1-1575-4b4c-b2fa-1c62db86bce1-config-volume\") pod \"915071d1-1575-4b4c-b2fa-1c62db86bce1\" (UID: \"915071d1-1575-4b4c-b2fa-1c62db86bce1\") " Mar 17 10:00:03 crc kubenswrapper[4813]: I0317 10:00:03.549148 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/915071d1-1575-4b4c-b2fa-1c62db86bce1-config-volume" (OuterVolumeSpecName: "config-volume") pod "915071d1-1575-4b4c-b2fa-1c62db86bce1" (UID: "915071d1-1575-4b4c-b2fa-1c62db86bce1"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 10:00:03 crc kubenswrapper[4813]: I0317 10:00:03.557751 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/915071d1-1575-4b4c-b2fa-1c62db86bce1-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "915071d1-1575-4b4c-b2fa-1c62db86bce1" (UID: "915071d1-1575-4b4c-b2fa-1c62db86bce1"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 10:00:03 crc kubenswrapper[4813]: I0317 10:00:03.557869 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/915071d1-1575-4b4c-b2fa-1c62db86bce1-kube-api-access-fks78" (OuterVolumeSpecName: "kube-api-access-fks78") pod "915071d1-1575-4b4c-b2fa-1c62db86bce1" (UID: "915071d1-1575-4b4c-b2fa-1c62db86bce1"). InnerVolumeSpecName "kube-api-access-fks78". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:00:03 crc kubenswrapper[4813]: I0317 10:00:03.650611 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fks78\" (UniqueName: \"kubernetes.io/projected/915071d1-1575-4b4c-b2fa-1c62db86bce1-kube-api-access-fks78\") on node \"crc\" DevicePath \"\"" Mar 17 10:00:03 crc kubenswrapper[4813]: I0317 10:00:03.650645 4813 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/915071d1-1575-4b4c-b2fa-1c62db86bce1-config-volume\") on node \"crc\" DevicePath \"\"" Mar 17 10:00:03 crc kubenswrapper[4813]: I0317 10:00:03.650655 4813 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/915071d1-1575-4b4c-b2fa-1c62db86bce1-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 17 10:00:03 crc kubenswrapper[4813]: I0317 10:00:03.924005 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562360-zm5lr" event={"ID":"915071d1-1575-4b4c-b2fa-1c62db86bce1","Type":"ContainerDied","Data":"302f0edcc364ff2e7c5589c43be6fe5f48197111b766549f0aa0abc7712ceb78"} Mar 17 10:00:03 crc kubenswrapper[4813]: I0317 10:00:03.924343 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="302f0edcc364ff2e7c5589c43be6fe5f48197111b766549f0aa0abc7712ceb78" Mar 17 10:00:03 crc kubenswrapper[4813]: I0317 10:00:03.924085 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562360-zm5lr" Mar 17 10:00:04 crc kubenswrapper[4813]: I0317 10:00:04.525801 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562315-j9xc6"] Mar 17 10:00:04 crc kubenswrapper[4813]: I0317 10:00:04.538475 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562315-j9xc6"] Mar 17 10:00:04 crc kubenswrapper[4813]: I0317 10:00:04.742855 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f4060f9-13f7-4b31-bc96-db1d3698de1a" path="/var/lib/kubelet/pods/3f4060f9-13f7-4b31-bc96-db1d3698de1a/volumes" Mar 17 10:00:04 crc kubenswrapper[4813]: I0317 10:00:04.932130 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562360-k9p9v" event={"ID":"0864972d-362c-4157-a956-5b99452985ab","Type":"ContainerStarted","Data":"6ac6b11a96e5e3aa979899ef2785d10eb5430beb4a593d579c6b58be9fe31f29"} Mar 17 10:00:04 crc kubenswrapper[4813]: I0317 10:00:04.956144 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562360-k9p9v" podStartSLOduration=1.733307762 podStartE2EDuration="4.956123431s" podCreationTimestamp="2026-03-17 10:00:00 +0000 UTC" firstStartedPulling="2026-03-17 10:00:01.13090837 +0000 UTC m=+3023.231711869" lastFinishedPulling="2026-03-17 10:00:04.353724039 +0000 UTC m=+3026.454527538" observedRunningTime="2026-03-17 10:00:04.946056228 +0000 UTC m=+3027.046859727" watchObservedRunningTime="2026-03-17 10:00:04.956123431 +0000 UTC m=+3027.056926930" Mar 17 10:00:05 crc kubenswrapper[4813]: I0317 10:00:05.942235 4813 generic.go:334] "Generic (PLEG): container finished" podID="0864972d-362c-4157-a956-5b99452985ab" containerID="6ac6b11a96e5e3aa979899ef2785d10eb5430beb4a593d579c6b58be9fe31f29" exitCode=0 Mar 17 10:00:05 crc kubenswrapper[4813]: I0317 10:00:05.942306 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562360-k9p9v" event={"ID":"0864972d-362c-4157-a956-5b99452985ab","Type":"ContainerDied","Data":"6ac6b11a96e5e3aa979899ef2785d10eb5430beb4a593d579c6b58be9fe31f29"} Mar 17 10:00:07 crc kubenswrapper[4813]: I0317 10:00:07.365836 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562360-k9p9v" Mar 17 10:00:07 crc kubenswrapper[4813]: I0317 10:00:07.538055 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tkmq7\" (UniqueName: \"kubernetes.io/projected/0864972d-362c-4157-a956-5b99452985ab-kube-api-access-tkmq7\") pod \"0864972d-362c-4157-a956-5b99452985ab\" (UID: \"0864972d-362c-4157-a956-5b99452985ab\") " Mar 17 10:00:07 crc kubenswrapper[4813]: I0317 10:00:07.581421 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0864972d-362c-4157-a956-5b99452985ab-kube-api-access-tkmq7" (OuterVolumeSpecName: "kube-api-access-tkmq7") pod "0864972d-362c-4157-a956-5b99452985ab" (UID: "0864972d-362c-4157-a956-5b99452985ab"). InnerVolumeSpecName "kube-api-access-tkmq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:00:07 crc kubenswrapper[4813]: I0317 10:00:07.640263 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tkmq7\" (UniqueName: \"kubernetes.io/projected/0864972d-362c-4157-a956-5b99452985ab-kube-api-access-tkmq7\") on node \"crc\" DevicePath \"\"" Mar 17 10:00:07 crc kubenswrapper[4813]: I0317 10:00:07.968313 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562360-k9p9v" event={"ID":"0864972d-362c-4157-a956-5b99452985ab","Type":"ContainerDied","Data":"2150fe11a9cae4d84d4e36bf76b15b93bcd77c2e2bd139988ce28ecd81259abe"} Mar 17 10:00:07 crc kubenswrapper[4813]: I0317 10:00:07.968390 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2150fe11a9cae4d84d4e36bf76b15b93bcd77c2e2bd139988ce28ecd81259abe" Mar 17 10:00:07 crc kubenswrapper[4813]: I0317 10:00:07.968402 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562360-k9p9v" Mar 17 10:00:08 crc kubenswrapper[4813]: I0317 10:00:08.063791 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562354-28hkn"] Mar 17 10:00:08 crc kubenswrapper[4813]: I0317 10:00:08.074629 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562354-28hkn"] Mar 17 10:00:08 crc kubenswrapper[4813]: I0317 10:00:08.745660 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:00:08 crc kubenswrapper[4813]: E0317 10:00:08.746731 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:00:08 crc kubenswrapper[4813]: I0317 10:00:08.749227 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5b02c90-782e-4dd8-81e7-44e868a1eb64" path="/var/lib/kubelet/pods/b5b02c90-782e-4dd8-81e7-44e868a1eb64/volumes" Mar 17 10:00:21 crc kubenswrapper[4813]: I0317 10:00:21.731745 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:00:21 crc kubenswrapper[4813]: E0317 10:00:21.734164 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:00:36 crc kubenswrapper[4813]: I0317 10:00:36.731435 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:00:36 crc kubenswrapper[4813]: E0317 10:00:36.732116 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:00:49 crc kubenswrapper[4813]: I0317 10:00:49.855529 4813 scope.go:117] "RemoveContainer" containerID="2fa5a95242f0483f2a8d0e8c9b8d64f9d9219802e44bd5d4f691be1ed406fe37" Mar 17 10:00:49 crc kubenswrapper[4813]: I0317 10:00:49.890380 4813 scope.go:117] "RemoveContainer" containerID="5460beadf7cd3ead87ea493ab4e4d67b338c964a2b007e8b646f9f96551be425" Mar 17 10:00:51 crc kubenswrapper[4813]: I0317 10:00:51.731305 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:00:51 crc kubenswrapper[4813]: E0317 10:00:51.731900 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:01:00 crc kubenswrapper[4813]: I0317 10:01:00.169203 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29562361-p6pcq"] Mar 17 10:01:00 crc kubenswrapper[4813]: E0317 10:01:00.171436 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="915071d1-1575-4b4c-b2fa-1c62db86bce1" containerName="collect-profiles" Mar 17 10:01:00 crc kubenswrapper[4813]: I0317 10:01:00.171469 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="915071d1-1575-4b4c-b2fa-1c62db86bce1" containerName="collect-profiles" Mar 17 10:01:00 crc kubenswrapper[4813]: E0317 10:01:00.171518 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0864972d-362c-4157-a956-5b99452985ab" containerName="oc" Mar 17 10:01:00 crc kubenswrapper[4813]: I0317 10:01:00.171534 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="0864972d-362c-4157-a956-5b99452985ab" containerName="oc" Mar 17 10:01:00 crc kubenswrapper[4813]: I0317 10:01:00.172258 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="0864972d-362c-4157-a956-5b99452985ab" containerName="oc" Mar 17 10:01:00 crc kubenswrapper[4813]: I0317 10:01:00.172299 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="915071d1-1575-4b4c-b2fa-1c62db86bce1" containerName="collect-profiles" Mar 17 10:01:00 crc kubenswrapper[4813]: I0317 10:01:00.174136 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29562361-p6pcq" Mar 17 10:01:00 crc kubenswrapper[4813]: I0317 10:01:00.179754 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29562361-p6pcq"] Mar 17 10:01:00 crc kubenswrapper[4813]: I0317 10:01:00.234928 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a4ef497-31ad-415a-8c23-1673c0e279cb-fernet-keys\") pod \"keystone-cron-29562361-p6pcq\" (UID: \"8a4ef497-31ad-415a-8c23-1673c0e279cb\") " pod="openstack/keystone-cron-29562361-p6pcq" Mar 17 10:01:00 crc kubenswrapper[4813]: I0317 10:01:00.235002 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a4ef497-31ad-415a-8c23-1673c0e279cb-config-data\") pod \"keystone-cron-29562361-p6pcq\" (UID: \"8a4ef497-31ad-415a-8c23-1673c0e279cb\") " pod="openstack/keystone-cron-29562361-p6pcq" Mar 17 10:01:00 crc kubenswrapper[4813]: I0317 10:01:00.235110 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a4ef497-31ad-415a-8c23-1673c0e279cb-combined-ca-bundle\") pod \"keystone-cron-29562361-p6pcq\" (UID: \"8a4ef497-31ad-415a-8c23-1673c0e279cb\") " pod="openstack/keystone-cron-29562361-p6pcq" Mar 17 10:01:00 crc kubenswrapper[4813]: I0317 10:01:00.235167 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8gkm\" (UniqueName: \"kubernetes.io/projected/8a4ef497-31ad-415a-8c23-1673c0e279cb-kube-api-access-b8gkm\") pod \"keystone-cron-29562361-p6pcq\" (UID: \"8a4ef497-31ad-415a-8c23-1673c0e279cb\") " pod="openstack/keystone-cron-29562361-p6pcq" Mar 17 10:01:00 crc kubenswrapper[4813]: I0317 10:01:00.337446 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a4ef497-31ad-415a-8c23-1673c0e279cb-combined-ca-bundle\") pod \"keystone-cron-29562361-p6pcq\" (UID: \"8a4ef497-31ad-415a-8c23-1673c0e279cb\") " pod="openstack/keystone-cron-29562361-p6pcq" Mar 17 10:01:00 crc kubenswrapper[4813]: I0317 10:01:00.338044 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8gkm\" (UniqueName: \"kubernetes.io/projected/8a4ef497-31ad-415a-8c23-1673c0e279cb-kube-api-access-b8gkm\") pod \"keystone-cron-29562361-p6pcq\" (UID: \"8a4ef497-31ad-415a-8c23-1673c0e279cb\") " pod="openstack/keystone-cron-29562361-p6pcq" Mar 17 10:01:00 crc kubenswrapper[4813]: I0317 10:01:00.338207 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a4ef497-31ad-415a-8c23-1673c0e279cb-fernet-keys\") pod \"keystone-cron-29562361-p6pcq\" (UID: \"8a4ef497-31ad-415a-8c23-1673c0e279cb\") " pod="openstack/keystone-cron-29562361-p6pcq" Mar 17 10:01:00 crc kubenswrapper[4813]: I0317 10:01:00.338320 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a4ef497-31ad-415a-8c23-1673c0e279cb-config-data\") pod \"keystone-cron-29562361-p6pcq\" (UID: \"8a4ef497-31ad-415a-8c23-1673c0e279cb\") " pod="openstack/keystone-cron-29562361-p6pcq" Mar 17 10:01:00 crc kubenswrapper[4813]: I0317 10:01:00.348286 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a4ef497-31ad-415a-8c23-1673c0e279cb-config-data\") pod \"keystone-cron-29562361-p6pcq\" (UID: \"8a4ef497-31ad-415a-8c23-1673c0e279cb\") " pod="openstack/keystone-cron-29562361-p6pcq" Mar 17 10:01:00 crc kubenswrapper[4813]: I0317 10:01:00.348877 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a4ef497-31ad-415a-8c23-1673c0e279cb-fernet-keys\") pod \"keystone-cron-29562361-p6pcq\" (UID: \"8a4ef497-31ad-415a-8c23-1673c0e279cb\") " pod="openstack/keystone-cron-29562361-p6pcq" Mar 17 10:01:00 crc kubenswrapper[4813]: I0317 10:01:00.348875 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a4ef497-31ad-415a-8c23-1673c0e279cb-combined-ca-bundle\") pod \"keystone-cron-29562361-p6pcq\" (UID: \"8a4ef497-31ad-415a-8c23-1673c0e279cb\") " pod="openstack/keystone-cron-29562361-p6pcq" Mar 17 10:01:00 crc kubenswrapper[4813]: I0317 10:01:00.365790 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8gkm\" (UniqueName: \"kubernetes.io/projected/8a4ef497-31ad-415a-8c23-1673c0e279cb-kube-api-access-b8gkm\") pod \"keystone-cron-29562361-p6pcq\" (UID: \"8a4ef497-31ad-415a-8c23-1673c0e279cb\") " pod="openstack/keystone-cron-29562361-p6pcq" Mar 17 10:01:00 crc kubenswrapper[4813]: I0317 10:01:00.499543 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29562361-p6pcq" Mar 17 10:01:00 crc kubenswrapper[4813]: I0317 10:01:00.931302 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29562361-p6pcq"] Mar 17 10:01:01 crc kubenswrapper[4813]: I0317 10:01:01.636544 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29562361-p6pcq" event={"ID":"8a4ef497-31ad-415a-8c23-1673c0e279cb","Type":"ContainerStarted","Data":"f466ed32d30068cb29fda435ec3f811e561e188a26410cde71b72475a6f3708b"} Mar 17 10:01:01 crc kubenswrapper[4813]: I0317 10:01:01.636904 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29562361-p6pcq" event={"ID":"8a4ef497-31ad-415a-8c23-1673c0e279cb","Type":"ContainerStarted","Data":"49021dce50ab519179c3a00c6e863e90229b7b8452f25c3ddadc2669e00bcc60"} Mar 17 10:01:01 crc kubenswrapper[4813]: I0317 10:01:01.662902 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29562361-p6pcq" podStartSLOduration=1.6628772029999999 podStartE2EDuration="1.662877203s" podCreationTimestamp="2026-03-17 10:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 10:01:01.6566769 +0000 UTC m=+3083.757480399" watchObservedRunningTime="2026-03-17 10:01:01.662877203 +0000 UTC m=+3083.763680702" Mar 17 10:01:03 crc kubenswrapper[4813]: I0317 10:01:03.731199 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:01:03 crc kubenswrapper[4813]: E0317 10:01:03.732380 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:01:04 crc kubenswrapper[4813]: I0317 10:01:04.672290 4813 generic.go:334] "Generic (PLEG): container finished" podID="8a4ef497-31ad-415a-8c23-1673c0e279cb" containerID="f466ed32d30068cb29fda435ec3f811e561e188a26410cde71b72475a6f3708b" exitCode=0 Mar 17 10:01:04 crc kubenswrapper[4813]: I0317 10:01:04.672381 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29562361-p6pcq" event={"ID":"8a4ef497-31ad-415a-8c23-1673c0e279cb","Type":"ContainerDied","Data":"f466ed32d30068cb29fda435ec3f811e561e188a26410cde71b72475a6f3708b"} Mar 17 10:01:06 crc kubenswrapper[4813]: I0317 10:01:06.152766 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29562361-p6pcq" Mar 17 10:01:06 crc kubenswrapper[4813]: I0317 10:01:06.270106 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a4ef497-31ad-415a-8c23-1673c0e279cb-fernet-keys\") pod \"8a4ef497-31ad-415a-8c23-1673c0e279cb\" (UID: \"8a4ef497-31ad-415a-8c23-1673c0e279cb\") " Mar 17 10:01:06 crc kubenswrapper[4813]: I0317 10:01:06.270353 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a4ef497-31ad-415a-8c23-1673c0e279cb-combined-ca-bundle\") pod \"8a4ef497-31ad-415a-8c23-1673c0e279cb\" (UID: \"8a4ef497-31ad-415a-8c23-1673c0e279cb\") " Mar 17 10:01:06 crc kubenswrapper[4813]: I0317 10:01:06.270454 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a4ef497-31ad-415a-8c23-1673c0e279cb-config-data\") pod \"8a4ef497-31ad-415a-8c23-1673c0e279cb\" (UID: \"8a4ef497-31ad-415a-8c23-1673c0e279cb\") " Mar 17 10:01:06 crc kubenswrapper[4813]: I0317 10:01:06.270685 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8gkm\" (UniqueName: \"kubernetes.io/projected/8a4ef497-31ad-415a-8c23-1673c0e279cb-kube-api-access-b8gkm\") pod \"8a4ef497-31ad-415a-8c23-1673c0e279cb\" (UID: \"8a4ef497-31ad-415a-8c23-1673c0e279cb\") " Mar 17 10:01:06 crc kubenswrapper[4813]: I0317 10:01:06.275799 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a4ef497-31ad-415a-8c23-1673c0e279cb-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "8a4ef497-31ad-415a-8c23-1673c0e279cb" (UID: "8a4ef497-31ad-415a-8c23-1673c0e279cb"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 10:01:06 crc kubenswrapper[4813]: I0317 10:01:06.276154 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a4ef497-31ad-415a-8c23-1673c0e279cb-kube-api-access-b8gkm" (OuterVolumeSpecName: "kube-api-access-b8gkm") pod "8a4ef497-31ad-415a-8c23-1673c0e279cb" (UID: "8a4ef497-31ad-415a-8c23-1673c0e279cb"). InnerVolumeSpecName "kube-api-access-b8gkm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:01:06 crc kubenswrapper[4813]: I0317 10:01:06.329341 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a4ef497-31ad-415a-8c23-1673c0e279cb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8a4ef497-31ad-415a-8c23-1673c0e279cb" (UID: "8a4ef497-31ad-415a-8c23-1673c0e279cb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 10:01:06 crc kubenswrapper[4813]: I0317 10:01:06.364774 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8a4ef497-31ad-415a-8c23-1673c0e279cb-config-data" (OuterVolumeSpecName: "config-data") pod "8a4ef497-31ad-415a-8c23-1673c0e279cb" (UID: "8a4ef497-31ad-415a-8c23-1673c0e279cb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 10:01:06 crc kubenswrapper[4813]: I0317 10:01:06.374546 4813 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8a4ef497-31ad-415a-8c23-1673c0e279cb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 17 10:01:06 crc kubenswrapper[4813]: I0317 10:01:06.374579 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8a4ef497-31ad-415a-8c23-1673c0e279cb-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 10:01:06 crc kubenswrapper[4813]: I0317 10:01:06.374591 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8gkm\" (UniqueName: \"kubernetes.io/projected/8a4ef497-31ad-415a-8c23-1673c0e279cb-kube-api-access-b8gkm\") on node \"crc\" DevicePath \"\"" Mar 17 10:01:06 crc kubenswrapper[4813]: I0317 10:01:06.374625 4813 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/8a4ef497-31ad-415a-8c23-1673c0e279cb-fernet-keys\") on node \"crc\" DevicePath \"\"" Mar 17 10:01:06 crc kubenswrapper[4813]: I0317 10:01:06.695784 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29562361-p6pcq" event={"ID":"8a4ef497-31ad-415a-8c23-1673c0e279cb","Type":"ContainerDied","Data":"49021dce50ab519179c3a00c6e863e90229b7b8452f25c3ddadc2669e00bcc60"} Mar 17 10:01:06 crc kubenswrapper[4813]: I0317 10:01:06.695851 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="49021dce50ab519179c3a00c6e863e90229b7b8452f25c3ddadc2669e00bcc60" Mar 17 10:01:06 crc kubenswrapper[4813]: I0317 10:01:06.695849 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29562361-p6pcq" Mar 17 10:01:17 crc kubenswrapper[4813]: I0317 10:01:17.730919 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:01:17 crc kubenswrapper[4813]: E0317 10:01:17.731690 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:01:32 crc kubenswrapper[4813]: I0317 10:01:32.732376 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:01:32 crc kubenswrapper[4813]: E0317 10:01:32.733824 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:01:44 crc kubenswrapper[4813]: I0317 10:01:44.753178 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:01:44 crc kubenswrapper[4813]: E0317 10:01:44.754475 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:01:56 crc kubenswrapper[4813]: I0317 10:01:56.731824 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:01:56 crc kubenswrapper[4813]: E0317 10:01:56.732933 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:02:00 crc kubenswrapper[4813]: I0317 10:02:00.159117 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562362-t4kqp"] Mar 17 10:02:00 crc kubenswrapper[4813]: E0317 10:02:00.160435 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a4ef497-31ad-415a-8c23-1673c0e279cb" containerName="keystone-cron" Mar 17 10:02:00 crc kubenswrapper[4813]: I0317 10:02:00.160454 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a4ef497-31ad-415a-8c23-1673c0e279cb" containerName="keystone-cron" Mar 17 10:02:00 crc kubenswrapper[4813]: I0317 10:02:00.160704 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a4ef497-31ad-415a-8c23-1673c0e279cb" containerName="keystone-cron" Mar 17 10:02:00 crc kubenswrapper[4813]: I0317 10:02:00.161498 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562362-t4kqp" Mar 17 10:02:00 crc kubenswrapper[4813]: I0317 10:02:00.165451 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 10:02:00 crc kubenswrapper[4813]: I0317 10:02:00.165453 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 10:02:00 crc kubenswrapper[4813]: I0317 10:02:00.165797 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 10:02:00 crc kubenswrapper[4813]: I0317 10:02:00.175557 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562362-t4kqp"] Mar 17 10:02:00 crc kubenswrapper[4813]: I0317 10:02:00.202908 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7lx7\" (UniqueName: \"kubernetes.io/projected/5050fcea-f024-4fda-9089-8596b42305f8-kube-api-access-z7lx7\") pod \"auto-csr-approver-29562362-t4kqp\" (UID: \"5050fcea-f024-4fda-9089-8596b42305f8\") " pod="openshift-infra/auto-csr-approver-29562362-t4kqp" Mar 17 10:02:00 crc kubenswrapper[4813]: I0317 10:02:00.306060 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7lx7\" (UniqueName: \"kubernetes.io/projected/5050fcea-f024-4fda-9089-8596b42305f8-kube-api-access-z7lx7\") pod \"auto-csr-approver-29562362-t4kqp\" (UID: \"5050fcea-f024-4fda-9089-8596b42305f8\") " pod="openshift-infra/auto-csr-approver-29562362-t4kqp" Mar 17 10:02:00 crc kubenswrapper[4813]: I0317 10:02:00.327450 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7lx7\" (UniqueName: \"kubernetes.io/projected/5050fcea-f024-4fda-9089-8596b42305f8-kube-api-access-z7lx7\") pod \"auto-csr-approver-29562362-t4kqp\" (UID: \"5050fcea-f024-4fda-9089-8596b42305f8\") " pod="openshift-infra/auto-csr-approver-29562362-t4kqp" Mar 17 10:02:00 crc kubenswrapper[4813]: I0317 10:02:00.498242 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562362-t4kqp" Mar 17 10:02:00 crc kubenswrapper[4813]: I0317 10:02:00.926020 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562362-t4kqp"] Mar 17 10:02:01 crc kubenswrapper[4813]: I0317 10:02:01.314362 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562362-t4kqp" event={"ID":"5050fcea-f024-4fda-9089-8596b42305f8","Type":"ContainerStarted","Data":"2095490c0028631649328ae2c1e6186c6763241fc846eca5a2c3dae9584ddbc4"} Mar 17 10:02:02 crc kubenswrapper[4813]: I0317 10:02:02.330347 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562362-t4kqp" event={"ID":"5050fcea-f024-4fda-9089-8596b42305f8","Type":"ContainerStarted","Data":"52a8a30c0cf59a3e466e6c4f24dd534a1f3933b6736975021fe13a668359ded6"} Mar 17 10:02:02 crc kubenswrapper[4813]: I0317 10:02:02.358594 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562362-t4kqp" podStartSLOduration=1.372694751 podStartE2EDuration="2.358567879s" podCreationTimestamp="2026-03-17 10:02:00 +0000 UTC" firstStartedPulling="2026-03-17 10:02:00.92985526 +0000 UTC m=+3143.030658759" lastFinishedPulling="2026-03-17 10:02:01.915728388 +0000 UTC m=+3144.016531887" observedRunningTime="2026-03-17 10:02:02.35187349 +0000 UTC m=+3144.452677029" watchObservedRunningTime="2026-03-17 10:02:02.358567879 +0000 UTC m=+3144.459371418" Mar 17 10:02:03 crc kubenswrapper[4813]: I0317 10:02:03.344560 4813 generic.go:334] "Generic (PLEG): container finished" podID="5050fcea-f024-4fda-9089-8596b42305f8" containerID="52a8a30c0cf59a3e466e6c4f24dd534a1f3933b6736975021fe13a668359ded6" exitCode=0 Mar 17 10:02:03 crc kubenswrapper[4813]: I0317 10:02:03.344676 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562362-t4kqp" event={"ID":"5050fcea-f024-4fda-9089-8596b42305f8","Type":"ContainerDied","Data":"52a8a30c0cf59a3e466e6c4f24dd534a1f3933b6736975021fe13a668359ded6"} Mar 17 10:02:04 crc kubenswrapper[4813]: I0317 10:02:04.863823 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562362-t4kqp" Mar 17 10:02:04 crc kubenswrapper[4813]: I0317 10:02:04.923283 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z7lx7\" (UniqueName: \"kubernetes.io/projected/5050fcea-f024-4fda-9089-8596b42305f8-kube-api-access-z7lx7\") pod \"5050fcea-f024-4fda-9089-8596b42305f8\" (UID: \"5050fcea-f024-4fda-9089-8596b42305f8\") " Mar 17 10:02:04 crc kubenswrapper[4813]: I0317 10:02:04.931464 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5050fcea-f024-4fda-9089-8596b42305f8-kube-api-access-z7lx7" (OuterVolumeSpecName: "kube-api-access-z7lx7") pod "5050fcea-f024-4fda-9089-8596b42305f8" (UID: "5050fcea-f024-4fda-9089-8596b42305f8"). InnerVolumeSpecName "kube-api-access-z7lx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:02:05 crc kubenswrapper[4813]: I0317 10:02:05.024830 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z7lx7\" (UniqueName: \"kubernetes.io/projected/5050fcea-f024-4fda-9089-8596b42305f8-kube-api-access-z7lx7\") on node \"crc\" DevicePath \"\"" Mar 17 10:02:05 crc kubenswrapper[4813]: I0317 10:02:05.367549 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562362-t4kqp" event={"ID":"5050fcea-f024-4fda-9089-8596b42305f8","Type":"ContainerDied","Data":"2095490c0028631649328ae2c1e6186c6763241fc846eca5a2c3dae9584ddbc4"} Mar 17 10:02:05 crc kubenswrapper[4813]: I0317 10:02:05.367996 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2095490c0028631649328ae2c1e6186c6763241fc846eca5a2c3dae9584ddbc4" Mar 17 10:02:05 crc kubenswrapper[4813]: I0317 10:02:05.367671 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562362-t4kqp" Mar 17 10:02:05 crc kubenswrapper[4813]: I0317 10:02:05.441920 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562356-kdgd8"] Mar 17 10:02:05 crc kubenswrapper[4813]: I0317 10:02:05.451429 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562356-kdgd8"] Mar 17 10:02:06 crc kubenswrapper[4813]: I0317 10:02:06.742271 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="069b5d80-8eea-46f2-b7e2-1843d75656d3" path="/var/lib/kubelet/pods/069b5d80-8eea-46f2-b7e2-1843d75656d3/volumes" Mar 17 10:02:09 crc kubenswrapper[4813]: I0317 10:02:09.731442 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:02:09 crc kubenswrapper[4813]: E0317 10:02:09.732399 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:02:24 crc kubenswrapper[4813]: I0317 10:02:24.730709 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:02:24 crc kubenswrapper[4813]: E0317 10:02:24.731403 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:02:38 crc kubenswrapper[4813]: I0317 10:02:38.742056 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:02:38 crc kubenswrapper[4813]: E0317 10:02:38.743030 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:02:49 crc kubenswrapper[4813]: I0317 10:02:49.742259 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:02:49 crc kubenswrapper[4813]: E0317 10:02:49.743484 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:02:50 crc kubenswrapper[4813]: I0317 10:02:50.104738 4813 scope.go:117] "RemoveContainer" containerID="f5b7f59921e77cafb74627590ca560bbc60aef6fbba2a9c48cc4a148af76c0b4" Mar 17 10:03:03 crc kubenswrapper[4813]: I0317 10:03:03.732395 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:03:03 crc kubenswrapper[4813]: E0317 10:03:03.732987 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:03:17 crc kubenswrapper[4813]: I0317 10:03:17.732582 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:03:17 crc kubenswrapper[4813]: E0317 10:03:17.734056 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:03:29 crc kubenswrapper[4813]: I0317 10:03:29.730461 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:03:29 crc kubenswrapper[4813]: E0317 10:03:29.732577 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:03:44 crc kubenswrapper[4813]: I0317 10:03:44.731045 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:03:44 crc kubenswrapper[4813]: E0317 10:03:44.731785 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:03:56 crc kubenswrapper[4813]: I0317 10:03:56.731027 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:03:56 crc kubenswrapper[4813]: E0317 10:03:56.733435 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:04:00 crc kubenswrapper[4813]: I0317 10:04:00.180445 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562364-z62gt"] Mar 17 10:04:00 crc kubenswrapper[4813]: E0317 10:04:00.181852 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5050fcea-f024-4fda-9089-8596b42305f8" containerName="oc" Mar 17 10:04:00 crc kubenswrapper[4813]: I0317 10:04:00.181868 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5050fcea-f024-4fda-9089-8596b42305f8" containerName="oc" Mar 17 10:04:00 crc kubenswrapper[4813]: I0317 10:04:00.182046 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5050fcea-f024-4fda-9089-8596b42305f8" containerName="oc" Mar 17 10:04:00 crc kubenswrapper[4813]: I0317 10:04:00.182764 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562364-z62gt" Mar 17 10:04:00 crc kubenswrapper[4813]: I0317 10:04:00.185855 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 10:04:00 crc kubenswrapper[4813]: I0317 10:04:00.186418 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 10:04:00 crc kubenswrapper[4813]: I0317 10:04:00.190288 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 10:04:00 crc kubenswrapper[4813]: I0317 10:04:00.199148 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562364-z62gt"] Mar 17 10:04:00 crc kubenswrapper[4813]: I0317 10:04:00.234883 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p57m9\" (UniqueName: \"kubernetes.io/projected/c5024d80-91a7-4953-9c8f-812e505b787c-kube-api-access-p57m9\") pod \"auto-csr-approver-29562364-z62gt\" (UID: \"c5024d80-91a7-4953-9c8f-812e505b787c\") " pod="openshift-infra/auto-csr-approver-29562364-z62gt" Mar 17 10:04:00 crc kubenswrapper[4813]: I0317 10:04:00.338233 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p57m9\" (UniqueName: \"kubernetes.io/projected/c5024d80-91a7-4953-9c8f-812e505b787c-kube-api-access-p57m9\") pod \"auto-csr-approver-29562364-z62gt\" (UID: \"c5024d80-91a7-4953-9c8f-812e505b787c\") " pod="openshift-infra/auto-csr-approver-29562364-z62gt" Mar 17 10:04:00 crc kubenswrapper[4813]: I0317 10:04:00.365235 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p57m9\" (UniqueName: \"kubernetes.io/projected/c5024d80-91a7-4953-9c8f-812e505b787c-kube-api-access-p57m9\") pod \"auto-csr-approver-29562364-z62gt\" (UID: \"c5024d80-91a7-4953-9c8f-812e505b787c\") " pod="openshift-infra/auto-csr-approver-29562364-z62gt" Mar 17 10:04:00 crc kubenswrapper[4813]: I0317 10:04:00.511584 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562364-z62gt" Mar 17 10:04:00 crc kubenswrapper[4813]: I0317 10:04:00.989332 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562364-z62gt"] Mar 17 10:04:01 crc kubenswrapper[4813]: I0317 10:04:01.000902 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 10:04:01 crc kubenswrapper[4813]: I0317 10:04:01.292892 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562364-z62gt" event={"ID":"c5024d80-91a7-4953-9c8f-812e505b787c","Type":"ContainerStarted","Data":"8e521db5907409da782b0831f879c12ff8eb3ea24e4e99374ef64d906da89a64"} Mar 17 10:04:02 crc kubenswrapper[4813]: I0317 10:04:02.306910 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562364-z62gt" event={"ID":"c5024d80-91a7-4953-9c8f-812e505b787c","Type":"ContainerStarted","Data":"82bb83955c3ed8ecebc47d512544710592de097b9756511eabf6f3c98f25f159"} Mar 17 10:04:02 crc kubenswrapper[4813]: I0317 10:04:02.340225 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562364-z62gt" podStartSLOduration=1.424499013 podStartE2EDuration="2.340201919s" podCreationTimestamp="2026-03-17 10:04:00 +0000 UTC" firstStartedPulling="2026-03-17 10:04:01.000697195 +0000 UTC m=+3263.101500684" lastFinishedPulling="2026-03-17 10:04:01.916400051 +0000 UTC m=+3264.017203590" observedRunningTime="2026-03-17 10:04:02.327298578 +0000 UTC m=+3264.428102097" watchObservedRunningTime="2026-03-17 10:04:02.340201919 +0000 UTC m=+3264.441005428" Mar 17 10:04:03 crc kubenswrapper[4813]: I0317 10:04:03.321369 4813 generic.go:334] "Generic (PLEG): container finished" podID="c5024d80-91a7-4953-9c8f-812e505b787c" containerID="82bb83955c3ed8ecebc47d512544710592de097b9756511eabf6f3c98f25f159" exitCode=0 Mar 17 10:04:03 crc kubenswrapper[4813]: I0317 10:04:03.321420 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562364-z62gt" event={"ID":"c5024d80-91a7-4953-9c8f-812e505b787c","Type":"ContainerDied","Data":"82bb83955c3ed8ecebc47d512544710592de097b9756511eabf6f3c98f25f159"} Mar 17 10:04:04 crc kubenswrapper[4813]: I0317 10:04:04.844781 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562364-z62gt" Mar 17 10:04:04 crc kubenswrapper[4813]: I0317 10:04:04.947939 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p57m9\" (UniqueName: \"kubernetes.io/projected/c5024d80-91a7-4953-9c8f-812e505b787c-kube-api-access-p57m9\") pod \"c5024d80-91a7-4953-9c8f-812e505b787c\" (UID: \"c5024d80-91a7-4953-9c8f-812e505b787c\") " Mar 17 10:04:04 crc kubenswrapper[4813]: I0317 10:04:04.954033 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c5024d80-91a7-4953-9c8f-812e505b787c-kube-api-access-p57m9" (OuterVolumeSpecName: "kube-api-access-p57m9") pod "c5024d80-91a7-4953-9c8f-812e505b787c" (UID: "c5024d80-91a7-4953-9c8f-812e505b787c"). InnerVolumeSpecName "kube-api-access-p57m9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:04:05 crc kubenswrapper[4813]: I0317 10:04:05.050262 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p57m9\" (UniqueName: \"kubernetes.io/projected/c5024d80-91a7-4953-9c8f-812e505b787c-kube-api-access-p57m9\") on node \"crc\" DevicePath \"\"" Mar 17 10:04:05 crc kubenswrapper[4813]: I0317 10:04:05.339992 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562364-z62gt" event={"ID":"c5024d80-91a7-4953-9c8f-812e505b787c","Type":"ContainerDied","Data":"8e521db5907409da782b0831f879c12ff8eb3ea24e4e99374ef64d906da89a64"} Mar 17 10:04:05 crc kubenswrapper[4813]: I0317 10:04:05.340027 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8e521db5907409da782b0831f879c12ff8eb3ea24e4e99374ef64d906da89a64" Mar 17 10:04:05 crc kubenswrapper[4813]: I0317 10:04:05.340069 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562364-z62gt" Mar 17 10:04:05 crc kubenswrapper[4813]: I0317 10:04:05.411114 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562358-kdqtd"] Mar 17 10:04:05 crc kubenswrapper[4813]: I0317 10:04:05.422770 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562358-kdqtd"] Mar 17 10:04:06 crc kubenswrapper[4813]: I0317 10:04:06.746045 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="16fe11ef-273a-403d-8b23-80c3955a869f" path="/var/lib/kubelet/pods/16fe11ef-273a-403d-8b23-80c3955a869f/volumes" Mar 17 10:04:11 crc kubenswrapper[4813]: I0317 10:04:11.731131 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:04:11 crc kubenswrapper[4813]: E0317 10:04:11.732428 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:04:16 crc kubenswrapper[4813]: I0317 10:04:16.752855 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-5r8z6"] Mar 17 10:04:16 crc kubenswrapper[4813]: E0317 10:04:16.754098 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c5024d80-91a7-4953-9c8f-812e505b787c" containerName="oc" Mar 17 10:04:16 crc kubenswrapper[4813]: I0317 10:04:16.754116 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c5024d80-91a7-4953-9c8f-812e505b787c" containerName="oc" Mar 17 10:04:16 crc kubenswrapper[4813]: I0317 10:04:16.754388 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c5024d80-91a7-4953-9c8f-812e505b787c" containerName="oc" Mar 17 10:04:16 crc kubenswrapper[4813]: I0317 10:04:16.756052 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5r8z6" Mar 17 10:04:16 crc kubenswrapper[4813]: I0317 10:04:16.786120 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5r8z6"] Mar 17 10:04:16 crc kubenswrapper[4813]: I0317 10:04:16.907307 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac9394e4-a370-49c4-8ae7-781201fa43fb-utilities\") pod \"certified-operators-5r8z6\" (UID: \"ac9394e4-a370-49c4-8ae7-781201fa43fb\") " pod="openshift-marketplace/certified-operators-5r8z6" Mar 17 10:04:16 crc kubenswrapper[4813]: I0317 10:04:16.907368 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qctb\" (UniqueName: \"kubernetes.io/projected/ac9394e4-a370-49c4-8ae7-781201fa43fb-kube-api-access-7qctb\") pod \"certified-operators-5r8z6\" (UID: \"ac9394e4-a370-49c4-8ae7-781201fa43fb\") " pod="openshift-marketplace/certified-operators-5r8z6" Mar 17 10:04:16 crc kubenswrapper[4813]: I0317 10:04:16.907556 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac9394e4-a370-49c4-8ae7-781201fa43fb-catalog-content\") pod \"certified-operators-5r8z6\" (UID: \"ac9394e4-a370-49c4-8ae7-781201fa43fb\") " pod="openshift-marketplace/certified-operators-5r8z6" Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.009855 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac9394e4-a370-49c4-8ae7-781201fa43fb-catalog-content\") pod \"certified-operators-5r8z6\" (UID: \"ac9394e4-a370-49c4-8ae7-781201fa43fb\") " pod="openshift-marketplace/certified-operators-5r8z6" Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.010303 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac9394e4-a370-49c4-8ae7-781201fa43fb-utilities\") pod \"certified-operators-5r8z6\" (UID: \"ac9394e4-a370-49c4-8ae7-781201fa43fb\") " pod="openshift-marketplace/certified-operators-5r8z6" Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.010392 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qctb\" (UniqueName: \"kubernetes.io/projected/ac9394e4-a370-49c4-8ae7-781201fa43fb-kube-api-access-7qctb\") pod \"certified-operators-5r8z6\" (UID: \"ac9394e4-a370-49c4-8ae7-781201fa43fb\") " pod="openshift-marketplace/certified-operators-5r8z6" Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.010647 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac9394e4-a370-49c4-8ae7-781201fa43fb-catalog-content\") pod \"certified-operators-5r8z6\" (UID: \"ac9394e4-a370-49c4-8ae7-781201fa43fb\") " pod="openshift-marketplace/certified-operators-5r8z6" Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.012988 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac9394e4-a370-49c4-8ae7-781201fa43fb-utilities\") pod \"certified-operators-5r8z6\" (UID: \"ac9394e4-a370-49c4-8ae7-781201fa43fb\") " pod="openshift-marketplace/certified-operators-5r8z6" Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.033275 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qctb\" (UniqueName: \"kubernetes.io/projected/ac9394e4-a370-49c4-8ae7-781201fa43fb-kube-api-access-7qctb\") pod \"certified-operators-5r8z6\" (UID: \"ac9394e4-a370-49c4-8ae7-781201fa43fb\") " pod="openshift-marketplace/certified-operators-5r8z6" Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.080124 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5r8z6" Mar 17 10:04:17 crc kubenswrapper[4813]: W0317 10:04:17.363217 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac9394e4_a370_49c4_8ae7_781201fa43fb.slice/crio-5a6a2bbbce08e63c0fbd2b7c03683007d0e72bf7b25bcc6428ca9a30966e2482 WatchSource:0}: Error finding container 5a6a2bbbce08e63c0fbd2b7c03683007d0e72bf7b25bcc6428ca9a30966e2482: Status 404 returned error can't find the container with id 5a6a2bbbce08e63c0fbd2b7c03683007d0e72bf7b25bcc6428ca9a30966e2482 Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.363851 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-5r8z6"] Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.481923 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r8z6" event={"ID":"ac9394e4-a370-49c4-8ae7-781201fa43fb","Type":"ContainerStarted","Data":"5a6a2bbbce08e63c0fbd2b7c03683007d0e72bf7b25bcc6428ca9a30966e2482"} Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.544467 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xdjv7"] Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.550132 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xdjv7" Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.557816 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xdjv7"] Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.622307 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4544e50b-a61e-4f73-86c7-09fecb9653a3-catalog-content\") pod \"redhat-operators-xdjv7\" (UID: \"4544e50b-a61e-4f73-86c7-09fecb9653a3\") " pod="openshift-marketplace/redhat-operators-xdjv7" Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.622486 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4544e50b-a61e-4f73-86c7-09fecb9653a3-utilities\") pod \"redhat-operators-xdjv7\" (UID: \"4544e50b-a61e-4f73-86c7-09fecb9653a3\") " pod="openshift-marketplace/redhat-operators-xdjv7" Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.622530 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-425vd\" (UniqueName: \"kubernetes.io/projected/4544e50b-a61e-4f73-86c7-09fecb9653a3-kube-api-access-425vd\") pod \"redhat-operators-xdjv7\" (UID: \"4544e50b-a61e-4f73-86c7-09fecb9653a3\") " pod="openshift-marketplace/redhat-operators-xdjv7" Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.724569 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-425vd\" (UniqueName: \"kubernetes.io/projected/4544e50b-a61e-4f73-86c7-09fecb9653a3-kube-api-access-425vd\") pod \"redhat-operators-xdjv7\" (UID: \"4544e50b-a61e-4f73-86c7-09fecb9653a3\") " pod="openshift-marketplace/redhat-operators-xdjv7" Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.724739 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4544e50b-a61e-4f73-86c7-09fecb9653a3-catalog-content\") pod \"redhat-operators-xdjv7\" (UID: \"4544e50b-a61e-4f73-86c7-09fecb9653a3\") " pod="openshift-marketplace/redhat-operators-xdjv7" Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.724903 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4544e50b-a61e-4f73-86c7-09fecb9653a3-utilities\") pod \"redhat-operators-xdjv7\" (UID: \"4544e50b-a61e-4f73-86c7-09fecb9653a3\") " pod="openshift-marketplace/redhat-operators-xdjv7" Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.725296 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4544e50b-a61e-4f73-86c7-09fecb9653a3-catalog-content\") pod \"redhat-operators-xdjv7\" (UID: \"4544e50b-a61e-4f73-86c7-09fecb9653a3\") " pod="openshift-marketplace/redhat-operators-xdjv7" Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.725426 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4544e50b-a61e-4f73-86c7-09fecb9653a3-utilities\") pod \"redhat-operators-xdjv7\" (UID: \"4544e50b-a61e-4f73-86c7-09fecb9653a3\") " pod="openshift-marketplace/redhat-operators-xdjv7" Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.742802 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-425vd\" (UniqueName: \"kubernetes.io/projected/4544e50b-a61e-4f73-86c7-09fecb9653a3-kube-api-access-425vd\") pod \"redhat-operators-xdjv7\" (UID: \"4544e50b-a61e-4f73-86c7-09fecb9653a3\") " pod="openshift-marketplace/redhat-operators-xdjv7" Mar 17 10:04:17 crc kubenswrapper[4813]: I0317 10:04:17.875147 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xdjv7" Mar 17 10:04:18 crc kubenswrapper[4813]: I0317 10:04:18.320121 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xdjv7"] Mar 17 10:04:18 crc kubenswrapper[4813]: I0317 10:04:18.491470 4813 generic.go:334] "Generic (PLEG): container finished" podID="ac9394e4-a370-49c4-8ae7-781201fa43fb" containerID="2702fef4d2211aca201f8037d871ee3fb2839c9b596a303602be6f08d991e5a2" exitCode=0 Mar 17 10:04:18 crc kubenswrapper[4813]: I0317 10:04:18.491568 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r8z6" event={"ID":"ac9394e4-a370-49c4-8ae7-781201fa43fb","Type":"ContainerDied","Data":"2702fef4d2211aca201f8037d871ee3fb2839c9b596a303602be6f08d991e5a2"} Mar 17 10:04:18 crc kubenswrapper[4813]: I0317 10:04:18.500578 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xdjv7" event={"ID":"4544e50b-a61e-4f73-86c7-09fecb9653a3","Type":"ContainerStarted","Data":"e152c368da6db09d83c3995e3d403a33b6e7db915b18624e70cbea1f14f6b8ed"} Mar 17 10:04:18 crc kubenswrapper[4813]: E0317 10:04:18.659572 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4544e50b_a61e_4f73_86c7_09fecb9653a3.slice/crio-d852fffd0c06e05a381ab640aeb740d02ab5ca7f171450ddd297695289fba970.scope\": RecentStats: unable to find data in memory cache]" Mar 17 10:04:19 crc kubenswrapper[4813]: I0317 10:04:19.512926 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r8z6" event={"ID":"ac9394e4-a370-49c4-8ae7-781201fa43fb","Type":"ContainerStarted","Data":"9e86f01e69a2c235d8b9dc6710e5fb1bfb6bcf2a13698be494ae6cd9c28edf5f"} Mar 17 10:04:19 crc kubenswrapper[4813]: I0317 10:04:19.514723 4813 generic.go:334] "Generic (PLEG): container finished" podID="4544e50b-a61e-4f73-86c7-09fecb9653a3" containerID="d852fffd0c06e05a381ab640aeb740d02ab5ca7f171450ddd297695289fba970" exitCode=0 Mar 17 10:04:19 crc kubenswrapper[4813]: I0317 10:04:19.514788 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xdjv7" event={"ID":"4544e50b-a61e-4f73-86c7-09fecb9653a3","Type":"ContainerDied","Data":"d852fffd0c06e05a381ab640aeb740d02ab5ca7f171450ddd297695289fba970"} Mar 17 10:04:20 crc kubenswrapper[4813]: I0317 10:04:20.524112 4813 generic.go:334] "Generic (PLEG): container finished" podID="ac9394e4-a370-49c4-8ae7-781201fa43fb" containerID="9e86f01e69a2c235d8b9dc6710e5fb1bfb6bcf2a13698be494ae6cd9c28edf5f" exitCode=0 Mar 17 10:04:20 crc kubenswrapper[4813]: I0317 10:04:20.524321 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r8z6" event={"ID":"ac9394e4-a370-49c4-8ae7-781201fa43fb","Type":"ContainerDied","Data":"9e86f01e69a2c235d8b9dc6710e5fb1bfb6bcf2a13698be494ae6cd9c28edf5f"} Mar 17 10:04:21 crc kubenswrapper[4813]: I0317 10:04:21.535878 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r8z6" event={"ID":"ac9394e4-a370-49c4-8ae7-781201fa43fb","Type":"ContainerStarted","Data":"f8553024c755df1bab82a535c77448464cf17b6037dad22bf2365442b129d6ae"} Mar 17 10:04:21 crc kubenswrapper[4813]: I0317 10:04:21.568834 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-5r8z6" podStartSLOduration=3.027134208 podStartE2EDuration="5.568812036s" podCreationTimestamp="2026-03-17 10:04:16 +0000 UTC" firstStartedPulling="2026-03-17 10:04:18.492895735 +0000 UTC m=+3280.593699234" lastFinishedPulling="2026-03-17 10:04:21.034573563 +0000 UTC m=+3283.135377062" observedRunningTime="2026-03-17 10:04:21.549767234 +0000 UTC m=+3283.650570733" watchObservedRunningTime="2026-03-17 10:04:21.568812036 +0000 UTC m=+3283.669615525" Mar 17 10:04:22 crc kubenswrapper[4813]: I0317 10:04:22.544564 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xdjv7" event={"ID":"4544e50b-a61e-4f73-86c7-09fecb9653a3","Type":"ContainerStarted","Data":"63d7ff0058a2952c6d6e221496e6a8c5c200593e8fb4f3fa46d6f1b9531edd5c"} Mar 17 10:04:25 crc kubenswrapper[4813]: I0317 10:04:25.730954 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:04:25 crc kubenswrapper[4813]: E0317 10:04:25.731474 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:04:26 crc kubenswrapper[4813]: I0317 10:04:26.581041 4813 generic.go:334] "Generic (PLEG): container finished" podID="4544e50b-a61e-4f73-86c7-09fecb9653a3" containerID="63d7ff0058a2952c6d6e221496e6a8c5c200593e8fb4f3fa46d6f1b9531edd5c" exitCode=0 Mar 17 10:04:26 crc kubenswrapper[4813]: I0317 10:04:26.581120 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xdjv7" event={"ID":"4544e50b-a61e-4f73-86c7-09fecb9653a3","Type":"ContainerDied","Data":"63d7ff0058a2952c6d6e221496e6a8c5c200593e8fb4f3fa46d6f1b9531edd5c"} Mar 17 10:04:27 crc kubenswrapper[4813]: I0317 10:04:27.080822 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-5r8z6" Mar 17 10:04:27 crc kubenswrapper[4813]: I0317 10:04:27.080868 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-5r8z6" Mar 17 10:04:27 crc kubenswrapper[4813]: I0317 10:04:27.129936 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-5r8z6" Mar 17 10:04:27 crc kubenswrapper[4813]: I0317 10:04:27.593313 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xdjv7" event={"ID":"4544e50b-a61e-4f73-86c7-09fecb9653a3","Type":"ContainerStarted","Data":"11a589c75855445cfb1af3b205f8041266d31a85064445c309e1bde2a838a57a"} Mar 17 10:04:27 crc kubenswrapper[4813]: I0317 10:04:27.620268 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xdjv7" podStartSLOduration=2.90930818 podStartE2EDuration="10.620244775s" podCreationTimestamp="2026-03-17 10:04:17 +0000 UTC" firstStartedPulling="2026-03-17 10:04:19.516899768 +0000 UTC m=+3281.617703297" lastFinishedPulling="2026-03-17 10:04:27.227836383 +0000 UTC m=+3289.328639892" observedRunningTime="2026-03-17 10:04:27.61620778 +0000 UTC m=+3289.717011279" watchObservedRunningTime="2026-03-17 10:04:27.620244775 +0000 UTC m=+3289.721048284" Mar 17 10:04:27 crc kubenswrapper[4813]: I0317 10:04:27.657900 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-5r8z6" Mar 17 10:04:27 crc kubenswrapper[4813]: I0317 10:04:27.875869 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xdjv7" Mar 17 10:04:27 crc kubenswrapper[4813]: I0317 10:04:27.875929 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xdjv7" Mar 17 10:04:28 crc kubenswrapper[4813]: I0317 10:04:28.918802 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xdjv7" podUID="4544e50b-a61e-4f73-86c7-09fecb9653a3" containerName="registry-server" probeResult="failure" output=< Mar 17 10:04:28 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Mar 17 10:04:28 crc kubenswrapper[4813]: > Mar 17 10:04:28 crc kubenswrapper[4813]: I0317 10:04:28.938771 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5r8z6"] Mar 17 10:04:29 crc kubenswrapper[4813]: I0317 10:04:29.610507 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-5r8z6" podUID="ac9394e4-a370-49c4-8ae7-781201fa43fb" containerName="registry-server" containerID="cri-o://f8553024c755df1bab82a535c77448464cf17b6037dad22bf2365442b129d6ae" gracePeriod=2 Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.038470 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5r8z6" Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.166269 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac9394e4-a370-49c4-8ae7-781201fa43fb-utilities\") pod \"ac9394e4-a370-49c4-8ae7-781201fa43fb\" (UID: \"ac9394e4-a370-49c4-8ae7-781201fa43fb\") " Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.166366 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac9394e4-a370-49c4-8ae7-781201fa43fb-catalog-content\") pod \"ac9394e4-a370-49c4-8ae7-781201fa43fb\" (UID: \"ac9394e4-a370-49c4-8ae7-781201fa43fb\") " Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.166411 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qctb\" (UniqueName: \"kubernetes.io/projected/ac9394e4-a370-49c4-8ae7-781201fa43fb-kube-api-access-7qctb\") pod \"ac9394e4-a370-49c4-8ae7-781201fa43fb\" (UID: \"ac9394e4-a370-49c4-8ae7-781201fa43fb\") " Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.167012 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac9394e4-a370-49c4-8ae7-781201fa43fb-utilities" (OuterVolumeSpecName: "utilities") pod "ac9394e4-a370-49c4-8ae7-781201fa43fb" (UID: "ac9394e4-a370-49c4-8ae7-781201fa43fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.176155 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac9394e4-a370-49c4-8ae7-781201fa43fb-kube-api-access-7qctb" (OuterVolumeSpecName: "kube-api-access-7qctb") pod "ac9394e4-a370-49c4-8ae7-781201fa43fb" (UID: "ac9394e4-a370-49c4-8ae7-781201fa43fb"). InnerVolumeSpecName "kube-api-access-7qctb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.210383 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac9394e4-a370-49c4-8ae7-781201fa43fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac9394e4-a370-49c4-8ae7-781201fa43fb" (UID: "ac9394e4-a370-49c4-8ae7-781201fa43fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.268559 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac9394e4-a370-49c4-8ae7-781201fa43fb-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.268589 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac9394e4-a370-49c4-8ae7-781201fa43fb-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.268626 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qctb\" (UniqueName: \"kubernetes.io/projected/ac9394e4-a370-49c4-8ae7-781201fa43fb-kube-api-access-7qctb\") on node \"crc\" DevicePath \"\"" Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.623478 4813 generic.go:334] "Generic (PLEG): container finished" podID="ac9394e4-a370-49c4-8ae7-781201fa43fb" containerID="f8553024c755df1bab82a535c77448464cf17b6037dad22bf2365442b129d6ae" exitCode=0 Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.623528 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r8z6" event={"ID":"ac9394e4-a370-49c4-8ae7-781201fa43fb","Type":"ContainerDied","Data":"f8553024c755df1bab82a535c77448464cf17b6037dad22bf2365442b129d6ae"} Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.623590 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-5r8z6" event={"ID":"ac9394e4-a370-49c4-8ae7-781201fa43fb","Type":"ContainerDied","Data":"5a6a2bbbce08e63c0fbd2b7c03683007d0e72bf7b25bcc6428ca9a30966e2482"} Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.623642 4813 scope.go:117] "RemoveContainer" containerID="f8553024c755df1bab82a535c77448464cf17b6037dad22bf2365442b129d6ae" Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.623644 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-5r8z6" Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.649445 4813 scope.go:117] "RemoveContainer" containerID="9e86f01e69a2c235d8b9dc6710e5fb1bfb6bcf2a13698be494ae6cd9c28edf5f" Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.676007 4813 scope.go:117] "RemoveContainer" containerID="2702fef4d2211aca201f8037d871ee3fb2839c9b596a303602be6f08d991e5a2" Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.678026 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-5r8z6"] Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.685677 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-5r8z6"] Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.734034 4813 scope.go:117] "RemoveContainer" containerID="f8553024c755df1bab82a535c77448464cf17b6037dad22bf2365442b129d6ae" Mar 17 10:04:30 crc kubenswrapper[4813]: E0317 10:04:30.734527 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8553024c755df1bab82a535c77448464cf17b6037dad22bf2365442b129d6ae\": container with ID starting with f8553024c755df1bab82a535c77448464cf17b6037dad22bf2365442b129d6ae not found: ID does not exist" containerID="f8553024c755df1bab82a535c77448464cf17b6037dad22bf2365442b129d6ae" Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.734622 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8553024c755df1bab82a535c77448464cf17b6037dad22bf2365442b129d6ae"} err="failed to get container status \"f8553024c755df1bab82a535c77448464cf17b6037dad22bf2365442b129d6ae\": rpc error: code = NotFound desc = could not find container \"f8553024c755df1bab82a535c77448464cf17b6037dad22bf2365442b129d6ae\": container with ID starting with f8553024c755df1bab82a535c77448464cf17b6037dad22bf2365442b129d6ae not found: ID does not exist" Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.734654 4813 scope.go:117] "RemoveContainer" containerID="9e86f01e69a2c235d8b9dc6710e5fb1bfb6bcf2a13698be494ae6cd9c28edf5f" Mar 17 10:04:30 crc kubenswrapper[4813]: E0317 10:04:30.734968 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e86f01e69a2c235d8b9dc6710e5fb1bfb6bcf2a13698be494ae6cd9c28edf5f\": container with ID starting with 9e86f01e69a2c235d8b9dc6710e5fb1bfb6bcf2a13698be494ae6cd9c28edf5f not found: ID does not exist" containerID="9e86f01e69a2c235d8b9dc6710e5fb1bfb6bcf2a13698be494ae6cd9c28edf5f" Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.735006 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e86f01e69a2c235d8b9dc6710e5fb1bfb6bcf2a13698be494ae6cd9c28edf5f"} err="failed to get container status \"9e86f01e69a2c235d8b9dc6710e5fb1bfb6bcf2a13698be494ae6cd9c28edf5f\": rpc error: code = NotFound desc = could not find container \"9e86f01e69a2c235d8b9dc6710e5fb1bfb6bcf2a13698be494ae6cd9c28edf5f\": container with ID starting with 9e86f01e69a2c235d8b9dc6710e5fb1bfb6bcf2a13698be494ae6cd9c28edf5f not found: ID does not exist" Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.735033 4813 scope.go:117] "RemoveContainer" containerID="2702fef4d2211aca201f8037d871ee3fb2839c9b596a303602be6f08d991e5a2" Mar 17 10:04:30 crc kubenswrapper[4813]: E0317 10:04:30.735359 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2702fef4d2211aca201f8037d871ee3fb2839c9b596a303602be6f08d991e5a2\": container with ID starting with 2702fef4d2211aca201f8037d871ee3fb2839c9b596a303602be6f08d991e5a2 not found: ID does not exist" containerID="2702fef4d2211aca201f8037d871ee3fb2839c9b596a303602be6f08d991e5a2" Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.735391 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2702fef4d2211aca201f8037d871ee3fb2839c9b596a303602be6f08d991e5a2"} err="failed to get container status \"2702fef4d2211aca201f8037d871ee3fb2839c9b596a303602be6f08d991e5a2\": rpc error: code = NotFound desc = could not find container \"2702fef4d2211aca201f8037d871ee3fb2839c9b596a303602be6f08d991e5a2\": container with ID starting with 2702fef4d2211aca201f8037d871ee3fb2839c9b596a303602be6f08d991e5a2 not found: ID does not exist" Mar 17 10:04:30 crc kubenswrapper[4813]: I0317 10:04:30.740523 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac9394e4-a370-49c4-8ae7-781201fa43fb" path="/var/lib/kubelet/pods/ac9394e4-a370-49c4-8ae7-781201fa43fb/volumes" Mar 17 10:04:38 crc kubenswrapper[4813]: I0317 10:04:38.957140 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-xdjv7" podUID="4544e50b-a61e-4f73-86c7-09fecb9653a3" containerName="registry-server" probeResult="failure" output=< Mar 17 10:04:38 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Mar 17 10:04:38 crc kubenswrapper[4813]: > Mar 17 10:04:39 crc kubenswrapper[4813]: I0317 10:04:39.731696 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:04:39 crc kubenswrapper[4813]: E0317 10:04:39.732268 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:04:47 crc kubenswrapper[4813]: I0317 10:04:47.931582 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xdjv7" Mar 17 10:04:48 crc kubenswrapper[4813]: I0317 10:04:48.014566 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xdjv7" Mar 17 10:04:48 crc kubenswrapper[4813]: I0317 10:04:48.953375 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xdjv7"] Mar 17 10:04:49 crc kubenswrapper[4813]: I0317 10:04:49.808658 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xdjv7" podUID="4544e50b-a61e-4f73-86c7-09fecb9653a3" containerName="registry-server" containerID="cri-o://11a589c75855445cfb1af3b205f8041266d31a85064445c309e1bde2a838a57a" gracePeriod=2 Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.246833 4813 scope.go:117] "RemoveContainer" containerID="eca8ccfa622bf9a226d07a404218fad4fe269b2b795671d0799ed86a629334b5" Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.394856 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xdjv7" Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.482458 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-425vd\" (UniqueName: \"kubernetes.io/projected/4544e50b-a61e-4f73-86c7-09fecb9653a3-kube-api-access-425vd\") pod \"4544e50b-a61e-4f73-86c7-09fecb9653a3\" (UID: \"4544e50b-a61e-4f73-86c7-09fecb9653a3\") " Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.482504 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4544e50b-a61e-4f73-86c7-09fecb9653a3-catalog-content\") pod \"4544e50b-a61e-4f73-86c7-09fecb9653a3\" (UID: \"4544e50b-a61e-4f73-86c7-09fecb9653a3\") " Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.482561 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4544e50b-a61e-4f73-86c7-09fecb9653a3-utilities\") pod \"4544e50b-a61e-4f73-86c7-09fecb9653a3\" (UID: \"4544e50b-a61e-4f73-86c7-09fecb9653a3\") " Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.484096 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4544e50b-a61e-4f73-86c7-09fecb9653a3-utilities" (OuterVolumeSpecName: "utilities") pod "4544e50b-a61e-4f73-86c7-09fecb9653a3" (UID: "4544e50b-a61e-4f73-86c7-09fecb9653a3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.488929 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4544e50b-a61e-4f73-86c7-09fecb9653a3-kube-api-access-425vd" (OuterVolumeSpecName: "kube-api-access-425vd") pod "4544e50b-a61e-4f73-86c7-09fecb9653a3" (UID: "4544e50b-a61e-4f73-86c7-09fecb9653a3"). InnerVolumeSpecName "kube-api-access-425vd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.584562 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-425vd\" (UniqueName: \"kubernetes.io/projected/4544e50b-a61e-4f73-86c7-09fecb9653a3-kube-api-access-425vd\") on node \"crc\" DevicePath \"\"" Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.584640 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4544e50b-a61e-4f73-86c7-09fecb9653a3-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.664624 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4544e50b-a61e-4f73-86c7-09fecb9653a3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4544e50b-a61e-4f73-86c7-09fecb9653a3" (UID: "4544e50b-a61e-4f73-86c7-09fecb9653a3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.691915 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4544e50b-a61e-4f73-86c7-09fecb9653a3-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.820306 4813 generic.go:334] "Generic (PLEG): container finished" podID="4544e50b-a61e-4f73-86c7-09fecb9653a3" containerID="11a589c75855445cfb1af3b205f8041266d31a85064445c309e1bde2a838a57a" exitCode=0 Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.820361 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xdjv7" event={"ID":"4544e50b-a61e-4f73-86c7-09fecb9653a3","Type":"ContainerDied","Data":"11a589c75855445cfb1af3b205f8041266d31a85064445c309e1bde2a838a57a"} Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.820397 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xdjv7" event={"ID":"4544e50b-a61e-4f73-86c7-09fecb9653a3","Type":"ContainerDied","Data":"e152c368da6db09d83c3995e3d403a33b6e7db915b18624e70cbea1f14f6b8ed"} Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.820421 4813 scope.go:117] "RemoveContainer" containerID="11a589c75855445cfb1af3b205f8041266d31a85064445c309e1bde2a838a57a" Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.820449 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xdjv7" Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.844282 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xdjv7"] Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.848540 4813 scope.go:117] "RemoveContainer" containerID="63d7ff0058a2952c6d6e221496e6a8c5c200593e8fb4f3fa46d6f1b9531edd5c" Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.852306 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xdjv7"] Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.877827 4813 scope.go:117] "RemoveContainer" containerID="d852fffd0c06e05a381ab640aeb740d02ab5ca7f171450ddd297695289fba970" Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.897700 4813 scope.go:117] "RemoveContainer" containerID="11a589c75855445cfb1af3b205f8041266d31a85064445c309e1bde2a838a57a" Mar 17 10:04:50 crc kubenswrapper[4813]: E0317 10:04:50.898145 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11a589c75855445cfb1af3b205f8041266d31a85064445c309e1bde2a838a57a\": container with ID starting with 11a589c75855445cfb1af3b205f8041266d31a85064445c309e1bde2a838a57a not found: ID does not exist" containerID="11a589c75855445cfb1af3b205f8041266d31a85064445c309e1bde2a838a57a" Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.898182 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11a589c75855445cfb1af3b205f8041266d31a85064445c309e1bde2a838a57a"} err="failed to get container status \"11a589c75855445cfb1af3b205f8041266d31a85064445c309e1bde2a838a57a\": rpc error: code = NotFound desc = could not find container \"11a589c75855445cfb1af3b205f8041266d31a85064445c309e1bde2a838a57a\": container with ID starting with 11a589c75855445cfb1af3b205f8041266d31a85064445c309e1bde2a838a57a not found: ID does not exist" Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.898207 4813 scope.go:117] "RemoveContainer" containerID="63d7ff0058a2952c6d6e221496e6a8c5c200593e8fb4f3fa46d6f1b9531edd5c" Mar 17 10:04:50 crc kubenswrapper[4813]: E0317 10:04:50.898507 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"63d7ff0058a2952c6d6e221496e6a8c5c200593e8fb4f3fa46d6f1b9531edd5c\": container with ID starting with 63d7ff0058a2952c6d6e221496e6a8c5c200593e8fb4f3fa46d6f1b9531edd5c not found: ID does not exist" containerID="63d7ff0058a2952c6d6e221496e6a8c5c200593e8fb4f3fa46d6f1b9531edd5c" Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.898534 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"63d7ff0058a2952c6d6e221496e6a8c5c200593e8fb4f3fa46d6f1b9531edd5c"} err="failed to get container status \"63d7ff0058a2952c6d6e221496e6a8c5c200593e8fb4f3fa46d6f1b9531edd5c\": rpc error: code = NotFound desc = could not find container \"63d7ff0058a2952c6d6e221496e6a8c5c200593e8fb4f3fa46d6f1b9531edd5c\": container with ID starting with 63d7ff0058a2952c6d6e221496e6a8c5c200593e8fb4f3fa46d6f1b9531edd5c not found: ID does not exist" Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.898556 4813 scope.go:117] "RemoveContainer" containerID="d852fffd0c06e05a381ab640aeb740d02ab5ca7f171450ddd297695289fba970" Mar 17 10:04:50 crc kubenswrapper[4813]: E0317 10:04:50.898831 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d852fffd0c06e05a381ab640aeb740d02ab5ca7f171450ddd297695289fba970\": container with ID starting with d852fffd0c06e05a381ab640aeb740d02ab5ca7f171450ddd297695289fba970 not found: ID does not exist" containerID="d852fffd0c06e05a381ab640aeb740d02ab5ca7f171450ddd297695289fba970" Mar 17 10:04:50 crc kubenswrapper[4813]: I0317 10:04:50.898856 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d852fffd0c06e05a381ab640aeb740d02ab5ca7f171450ddd297695289fba970"} err="failed to get container status \"d852fffd0c06e05a381ab640aeb740d02ab5ca7f171450ddd297695289fba970\": rpc error: code = NotFound desc = could not find container \"d852fffd0c06e05a381ab640aeb740d02ab5ca7f171450ddd297695289fba970\": container with ID starting with d852fffd0c06e05a381ab640aeb740d02ab5ca7f171450ddd297695289fba970 not found: ID does not exist" Mar 17 10:04:52 crc kubenswrapper[4813]: I0317 10:04:52.749696 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4544e50b-a61e-4f73-86c7-09fecb9653a3" path="/var/lib/kubelet/pods/4544e50b-a61e-4f73-86c7-09fecb9653a3/volumes" Mar 17 10:04:54 crc kubenswrapper[4813]: I0317 10:04:54.735272 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:04:55 crc kubenswrapper[4813]: I0317 10:04:55.879325 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerStarted","Data":"586f3d43249f84824ce70b39efe9e7b1dbced1e1e5bd5f1cc753140f0d665cb9"} Mar 17 10:06:00 crc kubenswrapper[4813]: I0317 10:06:00.150776 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562366-vpwp5"] Mar 17 10:06:00 crc kubenswrapper[4813]: E0317 10:06:00.151777 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4544e50b-a61e-4f73-86c7-09fecb9653a3" containerName="extract-utilities" Mar 17 10:06:00 crc kubenswrapper[4813]: I0317 10:06:00.151795 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4544e50b-a61e-4f73-86c7-09fecb9653a3" containerName="extract-utilities" Mar 17 10:06:00 crc kubenswrapper[4813]: E0317 10:06:00.151814 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4544e50b-a61e-4f73-86c7-09fecb9653a3" containerName="registry-server" Mar 17 10:06:00 crc kubenswrapper[4813]: I0317 10:06:00.151825 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4544e50b-a61e-4f73-86c7-09fecb9653a3" containerName="registry-server" Mar 17 10:06:00 crc kubenswrapper[4813]: E0317 10:06:00.151849 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac9394e4-a370-49c4-8ae7-781201fa43fb" containerName="registry-server" Mar 17 10:06:00 crc kubenswrapper[4813]: I0317 10:06:00.151858 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac9394e4-a370-49c4-8ae7-781201fa43fb" containerName="registry-server" Mar 17 10:06:00 crc kubenswrapper[4813]: E0317 10:06:00.151877 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac9394e4-a370-49c4-8ae7-781201fa43fb" containerName="extract-utilities" Mar 17 10:06:00 crc kubenswrapper[4813]: I0317 10:06:00.151885 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac9394e4-a370-49c4-8ae7-781201fa43fb" containerName="extract-utilities" Mar 17 10:06:00 crc kubenswrapper[4813]: E0317 10:06:00.151905 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4544e50b-a61e-4f73-86c7-09fecb9653a3" containerName="extract-content" Mar 17 10:06:00 crc kubenswrapper[4813]: I0317 10:06:00.151915 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="4544e50b-a61e-4f73-86c7-09fecb9653a3" containerName="extract-content" Mar 17 10:06:00 crc kubenswrapper[4813]: E0317 10:06:00.151927 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac9394e4-a370-49c4-8ae7-781201fa43fb" containerName="extract-content" Mar 17 10:06:00 crc kubenswrapper[4813]: I0317 10:06:00.151936 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac9394e4-a370-49c4-8ae7-781201fa43fb" containerName="extract-content" Mar 17 10:06:00 crc kubenswrapper[4813]: I0317 10:06:00.152141 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="4544e50b-a61e-4f73-86c7-09fecb9653a3" containerName="registry-server" Mar 17 10:06:00 crc kubenswrapper[4813]: I0317 10:06:00.152157 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac9394e4-a370-49c4-8ae7-781201fa43fb" containerName="registry-server" Mar 17 10:06:00 crc kubenswrapper[4813]: I0317 10:06:00.153013 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562366-vpwp5" Mar 17 10:06:00 crc kubenswrapper[4813]: I0317 10:06:00.156534 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 10:06:00 crc kubenswrapper[4813]: I0317 10:06:00.156737 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 10:06:00 crc kubenswrapper[4813]: I0317 10:06:00.157516 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 10:06:00 crc kubenswrapper[4813]: I0317 10:06:00.172120 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562366-vpwp5"] Mar 17 10:06:00 crc kubenswrapper[4813]: I0317 10:06:00.275363 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nhx86\" (UniqueName: \"kubernetes.io/projected/e04f629c-67d3-4578-bb16-cf3656db7e77-kube-api-access-nhx86\") pod \"auto-csr-approver-29562366-vpwp5\" (UID: \"e04f629c-67d3-4578-bb16-cf3656db7e77\") " pod="openshift-infra/auto-csr-approver-29562366-vpwp5" Mar 17 10:06:00 crc kubenswrapper[4813]: I0317 10:06:00.377468 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nhx86\" (UniqueName: \"kubernetes.io/projected/e04f629c-67d3-4578-bb16-cf3656db7e77-kube-api-access-nhx86\") pod \"auto-csr-approver-29562366-vpwp5\" (UID: \"e04f629c-67d3-4578-bb16-cf3656db7e77\") " pod="openshift-infra/auto-csr-approver-29562366-vpwp5" Mar 17 10:06:00 crc kubenswrapper[4813]: I0317 10:06:00.409339 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nhx86\" (UniqueName: \"kubernetes.io/projected/e04f629c-67d3-4578-bb16-cf3656db7e77-kube-api-access-nhx86\") pod \"auto-csr-approver-29562366-vpwp5\" (UID: \"e04f629c-67d3-4578-bb16-cf3656db7e77\") " pod="openshift-infra/auto-csr-approver-29562366-vpwp5" Mar 17 10:06:00 crc kubenswrapper[4813]: I0317 10:06:00.478895 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562366-vpwp5" Mar 17 10:06:00 crc kubenswrapper[4813]: I0317 10:06:00.926743 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562366-vpwp5"] Mar 17 10:06:01 crc kubenswrapper[4813]: I0317 10:06:01.585693 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562366-vpwp5" event={"ID":"e04f629c-67d3-4578-bb16-cf3656db7e77","Type":"ContainerStarted","Data":"7b6a95a735f37d2f3aadf6504c5d4b4d17d169e90a145e48c3bdf8dc5bca2c3f"} Mar 17 10:06:02 crc kubenswrapper[4813]: I0317 10:06:02.598548 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562366-vpwp5" event={"ID":"e04f629c-67d3-4578-bb16-cf3656db7e77","Type":"ContainerStarted","Data":"69fa314a70f55e1127d1a0ae14d1cd2f3fa8f93264d3f915c843376809f754c7"} Mar 17 10:06:02 crc kubenswrapper[4813]: I0317 10:06:02.623989 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562366-vpwp5" podStartSLOduration=1.4259415930000001 podStartE2EDuration="2.623955507s" podCreationTimestamp="2026-03-17 10:06:00 +0000 UTC" firstStartedPulling="2026-03-17 10:06:00.936527704 +0000 UTC m=+3383.037331203" lastFinishedPulling="2026-03-17 10:06:02.134541608 +0000 UTC m=+3384.235345117" observedRunningTime="2026-03-17 10:06:02.619236681 +0000 UTC m=+3384.720040220" watchObservedRunningTime="2026-03-17 10:06:02.623955507 +0000 UTC m=+3384.724759066" Mar 17 10:06:03 crc kubenswrapper[4813]: I0317 10:06:03.613994 4813 generic.go:334] "Generic (PLEG): container finished" podID="e04f629c-67d3-4578-bb16-cf3656db7e77" containerID="69fa314a70f55e1127d1a0ae14d1cd2f3fa8f93264d3f915c843376809f754c7" exitCode=0 Mar 17 10:06:03 crc kubenswrapper[4813]: I0317 10:06:03.614411 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562366-vpwp5" event={"ID":"e04f629c-67d3-4578-bb16-cf3656db7e77","Type":"ContainerDied","Data":"69fa314a70f55e1127d1a0ae14d1cd2f3fa8f93264d3f915c843376809f754c7"} Mar 17 10:06:05 crc kubenswrapper[4813]: I0317 10:06:05.162059 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562366-vpwp5" Mar 17 10:06:05 crc kubenswrapper[4813]: I0317 10:06:05.181540 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nhx86\" (UniqueName: \"kubernetes.io/projected/e04f629c-67d3-4578-bb16-cf3656db7e77-kube-api-access-nhx86\") pod \"e04f629c-67d3-4578-bb16-cf3656db7e77\" (UID: \"e04f629c-67d3-4578-bb16-cf3656db7e77\") " Mar 17 10:06:05 crc kubenswrapper[4813]: I0317 10:06:05.212676 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e04f629c-67d3-4578-bb16-cf3656db7e77-kube-api-access-nhx86" (OuterVolumeSpecName: "kube-api-access-nhx86") pod "e04f629c-67d3-4578-bb16-cf3656db7e77" (UID: "e04f629c-67d3-4578-bb16-cf3656db7e77"). InnerVolumeSpecName "kube-api-access-nhx86". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:06:05 crc kubenswrapper[4813]: I0317 10:06:05.284376 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nhx86\" (UniqueName: \"kubernetes.io/projected/e04f629c-67d3-4578-bb16-cf3656db7e77-kube-api-access-nhx86\") on node \"crc\" DevicePath \"\"" Mar 17 10:06:05 crc kubenswrapper[4813]: I0317 10:06:05.638744 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562366-vpwp5" event={"ID":"e04f629c-67d3-4578-bb16-cf3656db7e77","Type":"ContainerDied","Data":"7b6a95a735f37d2f3aadf6504c5d4b4d17d169e90a145e48c3bdf8dc5bca2c3f"} Mar 17 10:06:05 crc kubenswrapper[4813]: I0317 10:06:05.638829 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7b6a95a735f37d2f3aadf6504c5d4b4d17d169e90a145e48c3bdf8dc5bca2c3f" Mar 17 10:06:05 crc kubenswrapper[4813]: I0317 10:06:05.638909 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562366-vpwp5" Mar 17 10:06:05 crc kubenswrapper[4813]: I0317 10:06:05.721732 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562360-k9p9v"] Mar 17 10:06:05 crc kubenswrapper[4813]: I0317 10:06:05.735378 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562360-k9p9v"] Mar 17 10:06:06 crc kubenswrapper[4813]: I0317 10:06:06.752987 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0864972d-362c-4157-a956-5b99452985ab" path="/var/lib/kubelet/pods/0864972d-362c-4157-a956-5b99452985ab/volumes" Mar 17 10:06:43 crc kubenswrapper[4813]: I0317 10:06:43.803471 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gb6xt"] Mar 17 10:06:43 crc kubenswrapper[4813]: E0317 10:06:43.804744 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e04f629c-67d3-4578-bb16-cf3656db7e77" containerName="oc" Mar 17 10:06:43 crc kubenswrapper[4813]: I0317 10:06:43.804767 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e04f629c-67d3-4578-bb16-cf3656db7e77" containerName="oc" Mar 17 10:06:43 crc kubenswrapper[4813]: I0317 10:06:43.805124 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e04f629c-67d3-4578-bb16-cf3656db7e77" containerName="oc" Mar 17 10:06:43 crc kubenswrapper[4813]: I0317 10:06:43.810698 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gb6xt" Mar 17 10:06:43 crc kubenswrapper[4813]: I0317 10:06:43.821305 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gb6xt"] Mar 17 10:06:43 crc kubenswrapper[4813]: I0317 10:06:43.911352 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946-utilities\") pod \"community-operators-gb6xt\" (UID: \"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946\") " pod="openshift-marketplace/community-operators-gb6xt" Mar 17 10:06:43 crc kubenswrapper[4813]: I0317 10:06:43.911456 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nftv\" (UniqueName: \"kubernetes.io/projected/bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946-kube-api-access-2nftv\") pod \"community-operators-gb6xt\" (UID: \"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946\") " pod="openshift-marketplace/community-operators-gb6xt" Mar 17 10:06:43 crc kubenswrapper[4813]: I0317 10:06:43.911526 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946-catalog-content\") pod \"community-operators-gb6xt\" (UID: \"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946\") " pod="openshift-marketplace/community-operators-gb6xt" Mar 17 10:06:44 crc kubenswrapper[4813]: I0317 10:06:44.012869 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946-catalog-content\") pod \"community-operators-gb6xt\" (UID: \"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946\") " pod="openshift-marketplace/community-operators-gb6xt" Mar 17 10:06:44 crc kubenswrapper[4813]: I0317 10:06:44.013102 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946-utilities\") pod \"community-operators-gb6xt\" (UID: \"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946\") " pod="openshift-marketplace/community-operators-gb6xt" Mar 17 10:06:44 crc kubenswrapper[4813]: I0317 10:06:44.013189 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nftv\" (UniqueName: \"kubernetes.io/projected/bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946-kube-api-access-2nftv\") pod \"community-operators-gb6xt\" (UID: \"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946\") " pod="openshift-marketplace/community-operators-gb6xt" Mar 17 10:06:44 crc kubenswrapper[4813]: I0317 10:06:44.013418 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946-catalog-content\") pod \"community-operators-gb6xt\" (UID: \"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946\") " pod="openshift-marketplace/community-operators-gb6xt" Mar 17 10:06:44 crc kubenswrapper[4813]: I0317 10:06:44.013737 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946-utilities\") pod \"community-operators-gb6xt\" (UID: \"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946\") " pod="openshift-marketplace/community-operators-gb6xt" Mar 17 10:06:44 crc kubenswrapper[4813]: I0317 10:06:44.039057 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nftv\" (UniqueName: \"kubernetes.io/projected/bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946-kube-api-access-2nftv\") pod \"community-operators-gb6xt\" (UID: \"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946\") " pod="openshift-marketplace/community-operators-gb6xt" Mar 17 10:06:44 crc kubenswrapper[4813]: I0317 10:06:44.141153 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gb6xt" Mar 17 10:06:44 crc kubenswrapper[4813]: I0317 10:06:44.652870 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gb6xt"] Mar 17 10:06:44 crc kubenswrapper[4813]: W0317 10:06:44.659880 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbc59dc9e_0b8a_4f4e_b53e_cdcb89ece946.slice/crio-8d887ce7bf43ea476e13bd19fd59c21329cb43041e088bbbecac63ef28db70ca WatchSource:0}: Error finding container 8d887ce7bf43ea476e13bd19fd59c21329cb43041e088bbbecac63ef28db70ca: Status 404 returned error can't find the container with id 8d887ce7bf43ea476e13bd19fd59c21329cb43041e088bbbecac63ef28db70ca Mar 17 10:06:45 crc kubenswrapper[4813]: I0317 10:06:45.096169 4813 generic.go:334] "Generic (PLEG): container finished" podID="bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946" containerID="7153daf2c220d1a48a017c4731cb0068097f7a58dfe0d7fd214860da788ad338" exitCode=0 Mar 17 10:06:45 crc kubenswrapper[4813]: I0317 10:06:45.096249 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gb6xt" event={"ID":"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946","Type":"ContainerDied","Data":"7153daf2c220d1a48a017c4731cb0068097f7a58dfe0d7fd214860da788ad338"} Mar 17 10:06:45 crc kubenswrapper[4813]: I0317 10:06:45.096560 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gb6xt" event={"ID":"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946","Type":"ContainerStarted","Data":"8d887ce7bf43ea476e13bd19fd59c21329cb43041e088bbbecac63ef28db70ca"} Mar 17 10:06:46 crc kubenswrapper[4813]: I0317 10:06:46.106306 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gb6xt" event={"ID":"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946","Type":"ContainerStarted","Data":"a47376b6e9bc9b2f6508a1b8f70981cc5ce36e8692acdba98cadd064a07ff32d"} Mar 17 10:06:47 crc kubenswrapper[4813]: I0317 10:06:47.118740 4813 generic.go:334] "Generic (PLEG): container finished" podID="bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946" containerID="a47376b6e9bc9b2f6508a1b8f70981cc5ce36e8692acdba98cadd064a07ff32d" exitCode=0 Mar 17 10:06:47 crc kubenswrapper[4813]: I0317 10:06:47.119094 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gb6xt" event={"ID":"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946","Type":"ContainerDied","Data":"a47376b6e9bc9b2f6508a1b8f70981cc5ce36e8692acdba98cadd064a07ff32d"} Mar 17 10:06:49 crc kubenswrapper[4813]: I0317 10:06:49.138177 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gb6xt" event={"ID":"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946","Type":"ContainerStarted","Data":"8942f951d8e51f83881751e3f23e448648ff074f84c76c97646f2796b85d1676"} Mar 17 10:06:49 crc kubenswrapper[4813]: I0317 10:06:49.157110 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gb6xt" podStartSLOduration=3.290900985 podStartE2EDuration="6.157091294s" podCreationTimestamp="2026-03-17 10:06:43 +0000 UTC" firstStartedPulling="2026-03-17 10:06:45.098383512 +0000 UTC m=+3427.199187011" lastFinishedPulling="2026-03-17 10:06:47.964573791 +0000 UTC m=+3430.065377320" observedRunningTime="2026-03-17 10:06:49.156498516 +0000 UTC m=+3431.257302025" watchObservedRunningTime="2026-03-17 10:06:49.157091294 +0000 UTC m=+3431.257894793" Mar 17 10:06:50 crc kubenswrapper[4813]: I0317 10:06:50.455388 4813 scope.go:117] "RemoveContainer" containerID="6ac6b11a96e5e3aa979899ef2785d10eb5430beb4a593d579c6b58be9fe31f29" Mar 17 10:06:54 crc kubenswrapper[4813]: I0317 10:06:54.142232 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gb6xt" Mar 17 10:06:54 crc kubenswrapper[4813]: I0317 10:06:54.142701 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gb6xt" Mar 17 10:06:54 crc kubenswrapper[4813]: I0317 10:06:54.213993 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gb6xt" Mar 17 10:06:54 crc kubenswrapper[4813]: I0317 10:06:54.267490 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gb6xt" Mar 17 10:06:54 crc kubenswrapper[4813]: I0317 10:06:54.461456 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gb6xt"] Mar 17 10:06:56 crc kubenswrapper[4813]: I0317 10:06:56.222120 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gb6xt" podUID="bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946" containerName="registry-server" containerID="cri-o://8942f951d8e51f83881751e3f23e448648ff074f84c76c97646f2796b85d1676" gracePeriod=2 Mar 17 10:06:57 crc kubenswrapper[4813]: I0317 10:06:57.233225 4813 generic.go:334] "Generic (PLEG): container finished" podID="bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946" containerID="8942f951d8e51f83881751e3f23e448648ff074f84c76c97646f2796b85d1676" exitCode=0 Mar 17 10:06:57 crc kubenswrapper[4813]: I0317 10:06:57.233318 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gb6xt" event={"ID":"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946","Type":"ContainerDied","Data":"8942f951d8e51f83881751e3f23e448648ff074f84c76c97646f2796b85d1676"} Mar 17 10:06:57 crc kubenswrapper[4813]: I0317 10:06:57.414199 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gb6xt" Mar 17 10:06:57 crc kubenswrapper[4813]: I0317 10:06:57.524913 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946-catalog-content\") pod \"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946\" (UID: \"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946\") " Mar 17 10:06:57 crc kubenswrapper[4813]: I0317 10:06:57.525096 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946-utilities\") pod \"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946\" (UID: \"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946\") " Mar 17 10:06:57 crc kubenswrapper[4813]: I0317 10:06:57.525141 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nftv\" (UniqueName: \"kubernetes.io/projected/bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946-kube-api-access-2nftv\") pod \"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946\" (UID: \"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946\") " Mar 17 10:06:57 crc kubenswrapper[4813]: I0317 10:06:57.527332 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946-utilities" (OuterVolumeSpecName: "utilities") pod "bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946" (UID: "bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:06:57 crc kubenswrapper[4813]: I0317 10:06:57.532490 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946-kube-api-access-2nftv" (OuterVolumeSpecName: "kube-api-access-2nftv") pod "bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946" (UID: "bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946"). InnerVolumeSpecName "kube-api-access-2nftv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:06:57 crc kubenswrapper[4813]: I0317 10:06:57.577408 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946" (UID: "bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:06:57 crc kubenswrapper[4813]: I0317 10:06:57.627385 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 10:06:57 crc kubenswrapper[4813]: I0317 10:06:57.627435 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 10:06:57 crc kubenswrapper[4813]: I0317 10:06:57.627451 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nftv\" (UniqueName: \"kubernetes.io/projected/bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946-kube-api-access-2nftv\") on node \"crc\" DevicePath \"\"" Mar 17 10:06:58 crc kubenswrapper[4813]: I0317 10:06:58.258416 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gb6xt" event={"ID":"bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946","Type":"ContainerDied","Data":"8d887ce7bf43ea476e13bd19fd59c21329cb43041e088bbbecac63ef28db70ca"} Mar 17 10:06:58 crc kubenswrapper[4813]: I0317 10:06:58.258518 4813 scope.go:117] "RemoveContainer" containerID="8942f951d8e51f83881751e3f23e448648ff074f84c76c97646f2796b85d1676" Mar 17 10:06:58 crc kubenswrapper[4813]: I0317 10:06:58.258648 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gb6xt" Mar 17 10:06:58 crc kubenswrapper[4813]: I0317 10:06:58.306694 4813 scope.go:117] "RemoveContainer" containerID="a47376b6e9bc9b2f6508a1b8f70981cc5ce36e8692acdba98cadd064a07ff32d" Mar 17 10:06:58 crc kubenswrapper[4813]: I0317 10:06:58.325399 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gb6xt"] Mar 17 10:06:58 crc kubenswrapper[4813]: I0317 10:06:58.356501 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gb6xt"] Mar 17 10:06:58 crc kubenswrapper[4813]: I0317 10:06:58.357014 4813 scope.go:117] "RemoveContainer" containerID="7153daf2c220d1a48a017c4731cb0068097f7a58dfe0d7fd214860da788ad338" Mar 17 10:06:58 crc kubenswrapper[4813]: I0317 10:06:58.741353 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946" path="/var/lib/kubelet/pods/bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946/volumes" Mar 17 10:07:14 crc kubenswrapper[4813]: I0317 10:07:14.113961 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 10:07:14 crc kubenswrapper[4813]: I0317 10:07:14.115731 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 10:07:44 crc kubenswrapper[4813]: I0317 10:07:44.113702 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 10:07:44 crc kubenswrapper[4813]: I0317 10:07:44.115794 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 10:08:00 crc kubenswrapper[4813]: I0317 10:08:00.167799 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562368-tkpms"] Mar 17 10:08:00 crc kubenswrapper[4813]: E0317 10:08:00.168872 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946" containerName="extract-content" Mar 17 10:08:00 crc kubenswrapper[4813]: I0317 10:08:00.168897 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946" containerName="extract-content" Mar 17 10:08:00 crc kubenswrapper[4813]: E0317 10:08:00.168919 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946" containerName="registry-server" Mar 17 10:08:00 crc kubenswrapper[4813]: I0317 10:08:00.168930 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946" containerName="registry-server" Mar 17 10:08:00 crc kubenswrapper[4813]: E0317 10:08:00.168958 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946" containerName="extract-utilities" Mar 17 10:08:00 crc kubenswrapper[4813]: I0317 10:08:00.168967 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946" containerName="extract-utilities" Mar 17 10:08:00 crc kubenswrapper[4813]: I0317 10:08:00.169267 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc59dc9e-0b8a-4f4e-b53e-cdcb89ece946" containerName="registry-server" Mar 17 10:08:00 crc kubenswrapper[4813]: I0317 10:08:00.170151 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562368-tkpms" Mar 17 10:08:00 crc kubenswrapper[4813]: I0317 10:08:00.175320 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 10:08:00 crc kubenswrapper[4813]: I0317 10:08:00.176373 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 10:08:00 crc kubenswrapper[4813]: I0317 10:08:00.176698 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 10:08:00 crc kubenswrapper[4813]: I0317 10:08:00.190028 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562368-tkpms"] Mar 17 10:08:00 crc kubenswrapper[4813]: I0317 10:08:00.226942 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8b5c9\" (UniqueName: \"kubernetes.io/projected/39c26c40-ab53-43cf-93dd-14bd56d0dd6c-kube-api-access-8b5c9\") pod \"auto-csr-approver-29562368-tkpms\" (UID: \"39c26c40-ab53-43cf-93dd-14bd56d0dd6c\") " pod="openshift-infra/auto-csr-approver-29562368-tkpms" Mar 17 10:08:00 crc kubenswrapper[4813]: I0317 10:08:00.329204 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8b5c9\" (UniqueName: \"kubernetes.io/projected/39c26c40-ab53-43cf-93dd-14bd56d0dd6c-kube-api-access-8b5c9\") pod \"auto-csr-approver-29562368-tkpms\" (UID: \"39c26c40-ab53-43cf-93dd-14bd56d0dd6c\") " pod="openshift-infra/auto-csr-approver-29562368-tkpms" Mar 17 10:08:00 crc kubenswrapper[4813]: I0317 10:08:00.355990 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8b5c9\" (UniqueName: \"kubernetes.io/projected/39c26c40-ab53-43cf-93dd-14bd56d0dd6c-kube-api-access-8b5c9\") pod \"auto-csr-approver-29562368-tkpms\" (UID: \"39c26c40-ab53-43cf-93dd-14bd56d0dd6c\") " pod="openshift-infra/auto-csr-approver-29562368-tkpms" Mar 17 10:08:00 crc kubenswrapper[4813]: I0317 10:08:00.505592 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562368-tkpms" Mar 17 10:08:00 crc kubenswrapper[4813]: I0317 10:08:00.967945 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562368-tkpms"] Mar 17 10:08:01 crc kubenswrapper[4813]: I0317 10:08:01.968751 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562368-tkpms" event={"ID":"39c26c40-ab53-43cf-93dd-14bd56d0dd6c","Type":"ContainerStarted","Data":"36e2b58587df049d52cce3f6b181725a74f9ca9632c109755baf48ecbdc24b48"} Mar 17 10:08:02 crc kubenswrapper[4813]: I0317 10:08:02.981163 4813 generic.go:334] "Generic (PLEG): container finished" podID="39c26c40-ab53-43cf-93dd-14bd56d0dd6c" containerID="d5819ff4cd414054ad7e184c5b13093f822f73b7fea23519ebbf70e4503292a2" exitCode=0 Mar 17 10:08:02 crc kubenswrapper[4813]: I0317 10:08:02.981362 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562368-tkpms" event={"ID":"39c26c40-ab53-43cf-93dd-14bd56d0dd6c","Type":"ContainerDied","Data":"d5819ff4cd414054ad7e184c5b13093f822f73b7fea23519ebbf70e4503292a2"} Mar 17 10:08:04 crc kubenswrapper[4813]: I0317 10:08:04.492997 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562368-tkpms" Mar 17 10:08:04 crc kubenswrapper[4813]: I0317 10:08:04.621420 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8b5c9\" (UniqueName: \"kubernetes.io/projected/39c26c40-ab53-43cf-93dd-14bd56d0dd6c-kube-api-access-8b5c9\") pod \"39c26c40-ab53-43cf-93dd-14bd56d0dd6c\" (UID: \"39c26c40-ab53-43cf-93dd-14bd56d0dd6c\") " Mar 17 10:08:04 crc kubenswrapper[4813]: I0317 10:08:04.627491 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39c26c40-ab53-43cf-93dd-14bd56d0dd6c-kube-api-access-8b5c9" (OuterVolumeSpecName: "kube-api-access-8b5c9") pod "39c26c40-ab53-43cf-93dd-14bd56d0dd6c" (UID: "39c26c40-ab53-43cf-93dd-14bd56d0dd6c"). InnerVolumeSpecName "kube-api-access-8b5c9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:08:04 crc kubenswrapper[4813]: I0317 10:08:04.749306 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8b5c9\" (UniqueName: \"kubernetes.io/projected/39c26c40-ab53-43cf-93dd-14bd56d0dd6c-kube-api-access-8b5c9\") on node \"crc\" DevicePath \"\"" Mar 17 10:08:05 crc kubenswrapper[4813]: I0317 10:08:05.006350 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562368-tkpms" event={"ID":"39c26c40-ab53-43cf-93dd-14bd56d0dd6c","Type":"ContainerDied","Data":"36e2b58587df049d52cce3f6b181725a74f9ca9632c109755baf48ecbdc24b48"} Mar 17 10:08:05 crc kubenswrapper[4813]: I0317 10:08:05.006772 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="36e2b58587df049d52cce3f6b181725a74f9ca9632c109755baf48ecbdc24b48" Mar 17 10:08:05 crc kubenswrapper[4813]: I0317 10:08:05.006431 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562368-tkpms" Mar 17 10:08:05 crc kubenswrapper[4813]: I0317 10:08:05.596731 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562362-t4kqp"] Mar 17 10:08:05 crc kubenswrapper[4813]: I0317 10:08:05.615051 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562362-t4kqp"] Mar 17 10:08:06 crc kubenswrapper[4813]: I0317 10:08:06.746799 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5050fcea-f024-4fda-9089-8596b42305f8" path="/var/lib/kubelet/pods/5050fcea-f024-4fda-9089-8596b42305f8/volumes" Mar 17 10:08:14 crc kubenswrapper[4813]: I0317 10:08:14.113952 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 10:08:14 crc kubenswrapper[4813]: I0317 10:08:14.114576 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 10:08:14 crc kubenswrapper[4813]: I0317 10:08:14.114647 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 10:08:14 crc kubenswrapper[4813]: I0317 10:08:14.115427 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"586f3d43249f84824ce70b39efe9e7b1dbced1e1e5bd5f1cc753140f0d665cb9"} pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 10:08:14 crc kubenswrapper[4813]: I0317 10:08:14.115483 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" containerID="cri-o://586f3d43249f84824ce70b39efe9e7b1dbced1e1e5bd5f1cc753140f0d665cb9" gracePeriod=600 Mar 17 10:08:15 crc kubenswrapper[4813]: I0317 10:08:15.115179 4813 generic.go:334] "Generic (PLEG): container finished" podID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerID="586f3d43249f84824ce70b39efe9e7b1dbced1e1e5bd5f1cc753140f0d665cb9" exitCode=0 Mar 17 10:08:15 crc kubenswrapper[4813]: I0317 10:08:15.115252 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerDied","Data":"586f3d43249f84824ce70b39efe9e7b1dbced1e1e5bd5f1cc753140f0d665cb9"} Mar 17 10:08:15 crc kubenswrapper[4813]: I0317 10:08:15.115866 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerStarted","Data":"bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9"} Mar 17 10:08:15 crc kubenswrapper[4813]: I0317 10:08:15.115901 4813 scope.go:117] "RemoveContainer" containerID="9bae5bc22a37392885c391a42edbed5d1c23834757240b3a513244d76e5d7677" Mar 17 10:08:47 crc kubenswrapper[4813]: I0317 10:08:47.473463 4813 generic.go:334] "Generic (PLEG): container finished" podID="005f04b0-01f5-4bed-9d2c-2269ab67d27f" containerID="a645cf679eaec3380a711eb52ff7a30b787011ec69a129842da78db374bf1842" exitCode=0 Mar 17 10:08:47 crc kubenswrapper[4813]: I0317 10:08:47.473592 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"005f04b0-01f5-4bed-9d2c-2269ab67d27f","Type":"ContainerDied","Data":"a645cf679eaec3380a711eb52ff7a30b787011ec69a129842da78db374bf1842"} Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.006784 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.082400 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/005f04b0-01f5-4bed-9d2c-2269ab67d27f-ca-certs\") pod \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.082832 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/005f04b0-01f5-4bed-9d2c-2269ab67d27f-config-data\") pod \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.082864 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/005f04b0-01f5-4bed-9d2c-2269ab67d27f-test-operator-ephemeral-temporary\") pod \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.082923 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/005f04b0-01f5-4bed-9d2c-2269ab67d27f-openstack-config\") pod \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.082977 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/005f04b0-01f5-4bed-9d2c-2269ab67d27f-ssh-key\") pod \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.082996 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/005f04b0-01f5-4bed-9d2c-2269ab67d27f-test-operator-ephemeral-workdir\") pod \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.083041 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/005f04b0-01f5-4bed-9d2c-2269ab67d27f-openstack-config-secret\") pod \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.083059 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.083100 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwb74\" (UniqueName: \"kubernetes.io/projected/005f04b0-01f5-4bed-9d2c-2269ab67d27f-kube-api-access-fwb74\") pod \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\" (UID: \"005f04b0-01f5-4bed-9d2c-2269ab67d27f\") " Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.084634 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/005f04b0-01f5-4bed-9d2c-2269ab67d27f-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "005f04b0-01f5-4bed-9d2c-2269ab67d27f" (UID: "005f04b0-01f5-4bed-9d2c-2269ab67d27f"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.084689 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/005f04b0-01f5-4bed-9d2c-2269ab67d27f-config-data" (OuterVolumeSpecName: "config-data") pod "005f04b0-01f5-4bed-9d2c-2269ab67d27f" (UID: "005f04b0-01f5-4bed-9d2c-2269ab67d27f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.089093 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/005f04b0-01f5-4bed-9d2c-2269ab67d27f-kube-api-access-fwb74" (OuterVolumeSpecName: "kube-api-access-fwb74") pod "005f04b0-01f5-4bed-9d2c-2269ab67d27f" (UID: "005f04b0-01f5-4bed-9d2c-2269ab67d27f"). InnerVolumeSpecName "kube-api-access-fwb74". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.089201 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/005f04b0-01f5-4bed-9d2c-2269ab67d27f-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "005f04b0-01f5-4bed-9d2c-2269ab67d27f" (UID: "005f04b0-01f5-4bed-9d2c-2269ab67d27f"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.090389 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "test-operator-logs") pod "005f04b0-01f5-4bed-9d2c-2269ab67d27f" (UID: "005f04b0-01f5-4bed-9d2c-2269ab67d27f"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.114874 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/005f04b0-01f5-4bed-9d2c-2269ab67d27f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "005f04b0-01f5-4bed-9d2c-2269ab67d27f" (UID: "005f04b0-01f5-4bed-9d2c-2269ab67d27f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.118186 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/005f04b0-01f5-4bed-9d2c-2269ab67d27f-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "005f04b0-01f5-4bed-9d2c-2269ab67d27f" (UID: "005f04b0-01f5-4bed-9d2c-2269ab67d27f"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.128734 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/005f04b0-01f5-4bed-9d2c-2269ab67d27f-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "005f04b0-01f5-4bed-9d2c-2269ab67d27f" (UID: "005f04b0-01f5-4bed-9d2c-2269ab67d27f"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.163571 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/005f04b0-01f5-4bed-9d2c-2269ab67d27f-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "005f04b0-01f5-4bed-9d2c-2269ab67d27f" (UID: "005f04b0-01f5-4bed-9d2c-2269ab67d27f"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.186377 4813 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/005f04b0-01f5-4bed-9d2c-2269ab67d27f-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.186529 4813 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.186555 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwb74\" (UniqueName: \"kubernetes.io/projected/005f04b0-01f5-4bed-9d2c-2269ab67d27f-kube-api-access-fwb74\") on node \"crc\" DevicePath \"\"" Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.186575 4813 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/005f04b0-01f5-4bed-9d2c-2269ab67d27f-ca-certs\") on node \"crc\" DevicePath \"\"" Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.186595 4813 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/005f04b0-01f5-4bed-9d2c-2269ab67d27f-config-data\") on node \"crc\" DevicePath \"\"" Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.186641 4813 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/005f04b0-01f5-4bed-9d2c-2269ab67d27f-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.186754 4813 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/005f04b0-01f5-4bed-9d2c-2269ab67d27f-openstack-config\") on node \"crc\" DevicePath \"\"" Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.186836 4813 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/005f04b0-01f5-4bed-9d2c-2269ab67d27f-ssh-key\") on node \"crc\" DevicePath \"\"" Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.186855 4813 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/005f04b0-01f5-4bed-9d2c-2269ab67d27f-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.205891 4813 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.288359 4813 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.501492 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"005f04b0-01f5-4bed-9d2c-2269ab67d27f","Type":"ContainerDied","Data":"41d3487d7af3eae5152a5532d1c7b33b8d7a4e1b8abda035bb1321054aac8d7a"} Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.501559 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41d3487d7af3eae5152a5532d1c7b33b8d7a4e1b8abda035bb1321054aac8d7a" Mar 17 10:08:49 crc kubenswrapper[4813]: I0317 10:08:49.501589 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Mar 17 10:08:50 crc kubenswrapper[4813]: I0317 10:08:50.604419 4813 scope.go:117] "RemoveContainer" containerID="52a8a30c0cf59a3e466e6c4f24dd534a1f3933b6736975021fe13a668359ded6" Mar 17 10:08:51 crc kubenswrapper[4813]: I0317 10:08:51.715775 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 17 10:08:51 crc kubenswrapper[4813]: E0317 10:08:51.716456 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="005f04b0-01f5-4bed-9d2c-2269ab67d27f" containerName="tempest-tests-tempest-tests-runner" Mar 17 10:08:51 crc kubenswrapper[4813]: I0317 10:08:51.716481 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="005f04b0-01f5-4bed-9d2c-2269ab67d27f" containerName="tempest-tests-tempest-tests-runner" Mar 17 10:08:51 crc kubenswrapper[4813]: E0317 10:08:51.716517 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39c26c40-ab53-43cf-93dd-14bd56d0dd6c" containerName="oc" Mar 17 10:08:51 crc kubenswrapper[4813]: I0317 10:08:51.716531 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="39c26c40-ab53-43cf-93dd-14bd56d0dd6c" containerName="oc" Mar 17 10:08:51 crc kubenswrapper[4813]: I0317 10:08:51.716913 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="005f04b0-01f5-4bed-9d2c-2269ab67d27f" containerName="tempest-tests-tempest-tests-runner" Mar 17 10:08:51 crc kubenswrapper[4813]: I0317 10:08:51.716964 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="39c26c40-ab53-43cf-93dd-14bd56d0dd6c" containerName="oc" Mar 17 10:08:51 crc kubenswrapper[4813]: I0317 10:08:51.718077 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 17 10:08:51 crc kubenswrapper[4813]: I0317 10:08:51.721380 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-4lvrg" Mar 17 10:08:51 crc kubenswrapper[4813]: I0317 10:08:51.738581 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 17 10:08:51 crc kubenswrapper[4813]: I0317 10:08:51.842292 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5b58dbf4-d51d-4fe8-90d0-d295217084e1\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 17 10:08:51 crc kubenswrapper[4813]: I0317 10:08:51.842647 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8w6k\" (UniqueName: \"kubernetes.io/projected/5b58dbf4-d51d-4fe8-90d0-d295217084e1-kube-api-access-m8w6k\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5b58dbf4-d51d-4fe8-90d0-d295217084e1\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 17 10:08:51 crc kubenswrapper[4813]: I0317 10:08:51.944391 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5b58dbf4-d51d-4fe8-90d0-d295217084e1\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 17 10:08:51 crc kubenswrapper[4813]: I0317 10:08:51.944728 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8w6k\" (UniqueName: \"kubernetes.io/projected/5b58dbf4-d51d-4fe8-90d0-d295217084e1-kube-api-access-m8w6k\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5b58dbf4-d51d-4fe8-90d0-d295217084e1\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 17 10:08:51 crc kubenswrapper[4813]: I0317 10:08:51.945084 4813 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5b58dbf4-d51d-4fe8-90d0-d295217084e1\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 17 10:08:51 crc kubenswrapper[4813]: I0317 10:08:51.970972 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8w6k\" (UniqueName: \"kubernetes.io/projected/5b58dbf4-d51d-4fe8-90d0-d295217084e1-kube-api-access-m8w6k\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5b58dbf4-d51d-4fe8-90d0-d295217084e1\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 17 10:08:51 crc kubenswrapper[4813]: I0317 10:08:51.996097 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5b58dbf4-d51d-4fe8-90d0-d295217084e1\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 17 10:08:52 crc kubenswrapper[4813]: I0317 10:08:52.051384 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Mar 17 10:08:52 crc kubenswrapper[4813]: I0317 10:08:52.551100 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Mar 17 10:08:53 crc kubenswrapper[4813]: I0317 10:08:53.547011 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"5b58dbf4-d51d-4fe8-90d0-d295217084e1","Type":"ContainerStarted","Data":"e689a1598debc1921c2e798071575fd43ccc3d13e4e2f9dcef67085203730ca6"} Mar 17 10:08:54 crc kubenswrapper[4813]: I0317 10:08:54.564868 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"5b58dbf4-d51d-4fe8-90d0-d295217084e1","Type":"ContainerStarted","Data":"e2efde1e66fc8cdee4ffc93122fb6a8062726a9cc04199d5e50ff3bf4ba620d6"} Mar 17 10:08:54 crc kubenswrapper[4813]: I0317 10:08:54.589202 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.53458369 podStartE2EDuration="3.589174487s" podCreationTimestamp="2026-03-17 10:08:51 +0000 UTC" firstStartedPulling="2026-03-17 10:08:52.56463762 +0000 UTC m=+3554.665441119" lastFinishedPulling="2026-03-17 10:08:53.619228407 +0000 UTC m=+3555.720031916" observedRunningTime="2026-03-17 10:08:54.583094627 +0000 UTC m=+3556.683898166" watchObservedRunningTime="2026-03-17 10:08:54.589174487 +0000 UTC m=+3556.689978016" Mar 17 10:09:14 crc kubenswrapper[4813]: I0317 10:09:14.841316 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-d49m5/must-gather-676dh"] Mar 17 10:09:14 crc kubenswrapper[4813]: I0317 10:09:14.843359 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d49m5/must-gather-676dh" Mar 17 10:09:14 crc kubenswrapper[4813]: I0317 10:09:14.847269 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-d49m5"/"default-dockercfg-vzpgg" Mar 17 10:09:14 crc kubenswrapper[4813]: I0317 10:09:14.847437 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-d49m5"/"openshift-service-ca.crt" Mar 17 10:09:14 crc kubenswrapper[4813]: I0317 10:09:14.847997 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-d49m5"/"kube-root-ca.crt" Mar 17 10:09:14 crc kubenswrapper[4813]: I0317 10:09:14.863743 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-d49m5/must-gather-676dh"] Mar 17 10:09:14 crc kubenswrapper[4813]: I0317 10:09:14.926932 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9ed7b374-66e7-4391-9fa6-5d38d851d911-must-gather-output\") pod \"must-gather-676dh\" (UID: \"9ed7b374-66e7-4391-9fa6-5d38d851d911\") " pod="openshift-must-gather-d49m5/must-gather-676dh" Mar 17 10:09:14 crc kubenswrapper[4813]: I0317 10:09:14.927013 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zd6rs\" (UniqueName: \"kubernetes.io/projected/9ed7b374-66e7-4391-9fa6-5d38d851d911-kube-api-access-zd6rs\") pod \"must-gather-676dh\" (UID: \"9ed7b374-66e7-4391-9fa6-5d38d851d911\") " pod="openshift-must-gather-d49m5/must-gather-676dh" Mar 17 10:09:15 crc kubenswrapper[4813]: I0317 10:09:15.028792 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9ed7b374-66e7-4391-9fa6-5d38d851d911-must-gather-output\") pod \"must-gather-676dh\" (UID: \"9ed7b374-66e7-4391-9fa6-5d38d851d911\") " pod="openshift-must-gather-d49m5/must-gather-676dh" Mar 17 10:09:15 crc kubenswrapper[4813]: I0317 10:09:15.028849 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zd6rs\" (UniqueName: \"kubernetes.io/projected/9ed7b374-66e7-4391-9fa6-5d38d851d911-kube-api-access-zd6rs\") pod \"must-gather-676dh\" (UID: \"9ed7b374-66e7-4391-9fa6-5d38d851d911\") " pod="openshift-must-gather-d49m5/must-gather-676dh" Mar 17 10:09:15 crc kubenswrapper[4813]: I0317 10:09:15.029404 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9ed7b374-66e7-4391-9fa6-5d38d851d911-must-gather-output\") pod \"must-gather-676dh\" (UID: \"9ed7b374-66e7-4391-9fa6-5d38d851d911\") " pod="openshift-must-gather-d49m5/must-gather-676dh" Mar 17 10:09:15 crc kubenswrapper[4813]: I0317 10:09:15.045188 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zd6rs\" (UniqueName: \"kubernetes.io/projected/9ed7b374-66e7-4391-9fa6-5d38d851d911-kube-api-access-zd6rs\") pod \"must-gather-676dh\" (UID: \"9ed7b374-66e7-4391-9fa6-5d38d851d911\") " pod="openshift-must-gather-d49m5/must-gather-676dh" Mar 17 10:09:15 crc kubenswrapper[4813]: I0317 10:09:15.163027 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d49m5/must-gather-676dh" Mar 17 10:09:15 crc kubenswrapper[4813]: I0317 10:09:15.658663 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-d49m5/must-gather-676dh"] Mar 17 10:09:15 crc kubenswrapper[4813]: W0317 10:09:15.659124 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ed7b374_66e7_4391_9fa6_5d38d851d911.slice/crio-c929bf2f3a6751aa74ec937dbeca582be07a95286c53e3372d77b0e14f64ba77 WatchSource:0}: Error finding container c929bf2f3a6751aa74ec937dbeca582be07a95286c53e3372d77b0e14f64ba77: Status 404 returned error can't find the container with id c929bf2f3a6751aa74ec937dbeca582be07a95286c53e3372d77b0e14f64ba77 Mar 17 10:09:15 crc kubenswrapper[4813]: I0317 10:09:15.661736 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 10:09:15 crc kubenswrapper[4813]: I0317 10:09:15.804780 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d49m5/must-gather-676dh" event={"ID":"9ed7b374-66e7-4391-9fa6-5d38d851d911","Type":"ContainerStarted","Data":"c929bf2f3a6751aa74ec937dbeca582be07a95286c53e3372d77b0e14f64ba77"} Mar 17 10:09:22 crc kubenswrapper[4813]: I0317 10:09:22.881251 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d49m5/must-gather-676dh" event={"ID":"9ed7b374-66e7-4391-9fa6-5d38d851d911","Type":"ContainerStarted","Data":"274968342ec150566781c5e622b6db7f033fc8d0ac42a9dbf038fb70d308b6ce"} Mar 17 10:09:22 crc kubenswrapper[4813]: I0317 10:09:22.881848 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d49m5/must-gather-676dh" event={"ID":"9ed7b374-66e7-4391-9fa6-5d38d851d911","Type":"ContainerStarted","Data":"bd56a0dbbc97d78e263eec34ae6f377706150b613558a4d29ae98d4a3d9cbdda"} Mar 17 10:09:22 crc kubenswrapper[4813]: I0317 10:09:22.910884 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-d49m5/must-gather-676dh" podStartSLOduration=2.20661349 podStartE2EDuration="8.910865282s" podCreationTimestamp="2026-03-17 10:09:14 +0000 UTC" firstStartedPulling="2026-03-17 10:09:15.661281004 +0000 UTC m=+3577.762084533" lastFinishedPulling="2026-03-17 10:09:22.365532826 +0000 UTC m=+3584.466336325" observedRunningTime="2026-03-17 10:09:22.903846544 +0000 UTC m=+3585.004650083" watchObservedRunningTime="2026-03-17 10:09:22.910865282 +0000 UTC m=+3585.011668781" Mar 17 10:09:27 crc kubenswrapper[4813]: I0317 10:09:27.064364 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-d49m5/crc-debug-mmw8t"] Mar 17 10:09:27 crc kubenswrapper[4813]: I0317 10:09:27.066256 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d49m5/crc-debug-mmw8t" Mar 17 10:09:27 crc kubenswrapper[4813]: I0317 10:09:27.171522 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kts2\" (UniqueName: \"kubernetes.io/projected/68bcbff5-1452-47bf-9688-567614dd41f0-kube-api-access-9kts2\") pod \"crc-debug-mmw8t\" (UID: \"68bcbff5-1452-47bf-9688-567614dd41f0\") " pod="openshift-must-gather-d49m5/crc-debug-mmw8t" Mar 17 10:09:27 crc kubenswrapper[4813]: I0317 10:09:27.171638 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/68bcbff5-1452-47bf-9688-567614dd41f0-host\") pod \"crc-debug-mmw8t\" (UID: \"68bcbff5-1452-47bf-9688-567614dd41f0\") " pod="openshift-must-gather-d49m5/crc-debug-mmw8t" Mar 17 10:09:27 crc kubenswrapper[4813]: I0317 10:09:27.273605 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kts2\" (UniqueName: \"kubernetes.io/projected/68bcbff5-1452-47bf-9688-567614dd41f0-kube-api-access-9kts2\") pod \"crc-debug-mmw8t\" (UID: \"68bcbff5-1452-47bf-9688-567614dd41f0\") " pod="openshift-must-gather-d49m5/crc-debug-mmw8t" Mar 17 10:09:27 crc kubenswrapper[4813]: I0317 10:09:27.273958 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/68bcbff5-1452-47bf-9688-567614dd41f0-host\") pod \"crc-debug-mmw8t\" (UID: \"68bcbff5-1452-47bf-9688-567614dd41f0\") " pod="openshift-must-gather-d49m5/crc-debug-mmw8t" Mar 17 10:09:27 crc kubenswrapper[4813]: I0317 10:09:27.274148 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/68bcbff5-1452-47bf-9688-567614dd41f0-host\") pod \"crc-debug-mmw8t\" (UID: \"68bcbff5-1452-47bf-9688-567614dd41f0\") " pod="openshift-must-gather-d49m5/crc-debug-mmw8t" Mar 17 10:09:27 crc kubenswrapper[4813]: I0317 10:09:27.296475 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kts2\" (UniqueName: \"kubernetes.io/projected/68bcbff5-1452-47bf-9688-567614dd41f0-kube-api-access-9kts2\") pod \"crc-debug-mmw8t\" (UID: \"68bcbff5-1452-47bf-9688-567614dd41f0\") " pod="openshift-must-gather-d49m5/crc-debug-mmw8t" Mar 17 10:09:27 crc kubenswrapper[4813]: I0317 10:09:27.389230 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d49m5/crc-debug-mmw8t" Mar 17 10:09:27 crc kubenswrapper[4813]: W0317 10:09:27.435861 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod68bcbff5_1452_47bf_9688_567614dd41f0.slice/crio-affe9376880c74d7a35c421ee88d30f415334fc2b537d5ebafe005ffc0270514 WatchSource:0}: Error finding container affe9376880c74d7a35c421ee88d30f415334fc2b537d5ebafe005ffc0270514: Status 404 returned error can't find the container with id affe9376880c74d7a35c421ee88d30f415334fc2b537d5ebafe005ffc0270514 Mar 17 10:09:27 crc kubenswrapper[4813]: I0317 10:09:27.940516 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d49m5/crc-debug-mmw8t" event={"ID":"68bcbff5-1452-47bf-9688-567614dd41f0","Type":"ContainerStarted","Data":"affe9376880c74d7a35c421ee88d30f415334fc2b537d5ebafe005ffc0270514"} Mar 17 10:09:39 crc kubenswrapper[4813]: I0317 10:09:39.044509 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d49m5/crc-debug-mmw8t" event={"ID":"68bcbff5-1452-47bf-9688-567614dd41f0","Type":"ContainerStarted","Data":"07714735dd2134ad863dcb2368769844bf3220bfb769ca505a47f92572e07e2f"} Mar 17 10:09:39 crc kubenswrapper[4813]: I0317 10:09:39.064749 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-d49m5/crc-debug-mmw8t" podStartSLOduration=0.950271818 podStartE2EDuration="12.064731087s" podCreationTimestamp="2026-03-17 10:09:27 +0000 UTC" firstStartedPulling="2026-03-17 10:09:27.43790179 +0000 UTC m=+3589.538705289" lastFinishedPulling="2026-03-17 10:09:38.552361039 +0000 UTC m=+3600.653164558" observedRunningTime="2026-03-17 10:09:39.060669591 +0000 UTC m=+3601.161473090" watchObservedRunningTime="2026-03-17 10:09:39.064731087 +0000 UTC m=+3601.165534586" Mar 17 10:10:00 crc kubenswrapper[4813]: I0317 10:10:00.156142 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562370-8rqq9"] Mar 17 10:10:00 crc kubenswrapper[4813]: I0317 10:10:00.158332 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562370-8rqq9" Mar 17 10:10:00 crc kubenswrapper[4813]: I0317 10:10:00.160759 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 10:10:00 crc kubenswrapper[4813]: I0317 10:10:00.161894 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 10:10:00 crc kubenswrapper[4813]: I0317 10:10:00.162438 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 10:10:00 crc kubenswrapper[4813]: I0317 10:10:00.169283 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562370-8rqq9"] Mar 17 10:10:00 crc kubenswrapper[4813]: I0317 10:10:00.258845 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m4jt7\" (UniqueName: \"kubernetes.io/projected/a52ad15e-4714-47c1-87e7-b254fdd2d975-kube-api-access-m4jt7\") pod \"auto-csr-approver-29562370-8rqq9\" (UID: \"a52ad15e-4714-47c1-87e7-b254fdd2d975\") " pod="openshift-infra/auto-csr-approver-29562370-8rqq9" Mar 17 10:10:00 crc kubenswrapper[4813]: I0317 10:10:00.360822 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m4jt7\" (UniqueName: \"kubernetes.io/projected/a52ad15e-4714-47c1-87e7-b254fdd2d975-kube-api-access-m4jt7\") pod \"auto-csr-approver-29562370-8rqq9\" (UID: \"a52ad15e-4714-47c1-87e7-b254fdd2d975\") " pod="openshift-infra/auto-csr-approver-29562370-8rqq9" Mar 17 10:10:00 crc kubenswrapper[4813]: I0317 10:10:00.388394 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m4jt7\" (UniqueName: \"kubernetes.io/projected/a52ad15e-4714-47c1-87e7-b254fdd2d975-kube-api-access-m4jt7\") pod \"auto-csr-approver-29562370-8rqq9\" (UID: \"a52ad15e-4714-47c1-87e7-b254fdd2d975\") " pod="openshift-infra/auto-csr-approver-29562370-8rqq9" Mar 17 10:10:00 crc kubenswrapper[4813]: I0317 10:10:00.486309 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562370-8rqq9" Mar 17 10:10:00 crc kubenswrapper[4813]: I0317 10:10:00.960027 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562370-8rqq9"] Mar 17 10:10:01 crc kubenswrapper[4813]: I0317 10:10:01.237740 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562370-8rqq9" event={"ID":"a52ad15e-4714-47c1-87e7-b254fdd2d975","Type":"ContainerStarted","Data":"164c0fd42cdd9543f37ddc6c0791ff977dd5a9ef72cce90d976716d0d60bc0ec"} Mar 17 10:10:03 crc kubenswrapper[4813]: I0317 10:10:03.257235 4813 generic.go:334] "Generic (PLEG): container finished" podID="a52ad15e-4714-47c1-87e7-b254fdd2d975" containerID="03a45c4015d4a93a56227694dc938a468d8eebef6e657e2ff4bd8460561282c1" exitCode=0 Mar 17 10:10:03 crc kubenswrapper[4813]: I0317 10:10:03.257349 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562370-8rqq9" event={"ID":"a52ad15e-4714-47c1-87e7-b254fdd2d975","Type":"ContainerDied","Data":"03a45c4015d4a93a56227694dc938a468d8eebef6e657e2ff4bd8460561282c1"} Mar 17 10:10:04 crc kubenswrapper[4813]: I0317 10:10:04.668523 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562370-8rqq9" Mar 17 10:10:04 crc kubenswrapper[4813]: I0317 10:10:04.858435 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m4jt7\" (UniqueName: \"kubernetes.io/projected/a52ad15e-4714-47c1-87e7-b254fdd2d975-kube-api-access-m4jt7\") pod \"a52ad15e-4714-47c1-87e7-b254fdd2d975\" (UID: \"a52ad15e-4714-47c1-87e7-b254fdd2d975\") " Mar 17 10:10:04 crc kubenswrapper[4813]: I0317 10:10:04.870139 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a52ad15e-4714-47c1-87e7-b254fdd2d975-kube-api-access-m4jt7" (OuterVolumeSpecName: "kube-api-access-m4jt7") pod "a52ad15e-4714-47c1-87e7-b254fdd2d975" (UID: "a52ad15e-4714-47c1-87e7-b254fdd2d975"). InnerVolumeSpecName "kube-api-access-m4jt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:10:04 crc kubenswrapper[4813]: I0317 10:10:04.961025 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m4jt7\" (UniqueName: \"kubernetes.io/projected/a52ad15e-4714-47c1-87e7-b254fdd2d975-kube-api-access-m4jt7\") on node \"crc\" DevicePath \"\"" Mar 17 10:10:05 crc kubenswrapper[4813]: I0317 10:10:05.276779 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562370-8rqq9" event={"ID":"a52ad15e-4714-47c1-87e7-b254fdd2d975","Type":"ContainerDied","Data":"164c0fd42cdd9543f37ddc6c0791ff977dd5a9ef72cce90d976716d0d60bc0ec"} Mar 17 10:10:05 crc kubenswrapper[4813]: I0317 10:10:05.277050 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="164c0fd42cdd9543f37ddc6c0791ff977dd5a9ef72cce90d976716d0d60bc0ec" Mar 17 10:10:05 crc kubenswrapper[4813]: I0317 10:10:05.276826 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562370-8rqq9" Mar 17 10:10:05 crc kubenswrapper[4813]: I0317 10:10:05.739048 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562364-z62gt"] Mar 17 10:10:05 crc kubenswrapper[4813]: I0317 10:10:05.747514 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562364-z62gt"] Mar 17 10:10:06 crc kubenswrapper[4813]: I0317 10:10:06.743841 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c5024d80-91a7-4953-9c8f-812e505b787c" path="/var/lib/kubelet/pods/c5024d80-91a7-4953-9c8f-812e505b787c/volumes" Mar 17 10:10:07 crc kubenswrapper[4813]: I0317 10:10:07.961420 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-h44mk"] Mar 17 10:10:07 crc kubenswrapper[4813]: E0317 10:10:07.962253 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a52ad15e-4714-47c1-87e7-b254fdd2d975" containerName="oc" Mar 17 10:10:07 crc kubenswrapper[4813]: I0317 10:10:07.962272 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="a52ad15e-4714-47c1-87e7-b254fdd2d975" containerName="oc" Mar 17 10:10:07 crc kubenswrapper[4813]: I0317 10:10:07.962523 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="a52ad15e-4714-47c1-87e7-b254fdd2d975" containerName="oc" Mar 17 10:10:07 crc kubenswrapper[4813]: I0317 10:10:07.964160 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h44mk" Mar 17 10:10:07 crc kubenswrapper[4813]: I0317 10:10:07.978677 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h44mk"] Mar 17 10:10:08 crc kubenswrapper[4813]: I0317 10:10:08.016478 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5264d4a2-bda1-4cdd-a63e-e924bd43bc82-catalog-content\") pod \"redhat-marketplace-h44mk\" (UID: \"5264d4a2-bda1-4cdd-a63e-e924bd43bc82\") " pod="openshift-marketplace/redhat-marketplace-h44mk" Mar 17 10:10:08 crc kubenswrapper[4813]: I0317 10:10:08.016544 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fpr48\" (UniqueName: \"kubernetes.io/projected/5264d4a2-bda1-4cdd-a63e-e924bd43bc82-kube-api-access-fpr48\") pod \"redhat-marketplace-h44mk\" (UID: \"5264d4a2-bda1-4cdd-a63e-e924bd43bc82\") " pod="openshift-marketplace/redhat-marketplace-h44mk" Mar 17 10:10:08 crc kubenswrapper[4813]: I0317 10:10:08.016584 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5264d4a2-bda1-4cdd-a63e-e924bd43bc82-utilities\") pod \"redhat-marketplace-h44mk\" (UID: \"5264d4a2-bda1-4cdd-a63e-e924bd43bc82\") " pod="openshift-marketplace/redhat-marketplace-h44mk" Mar 17 10:10:08 crc kubenswrapper[4813]: I0317 10:10:08.117646 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5264d4a2-bda1-4cdd-a63e-e924bd43bc82-catalog-content\") pod \"redhat-marketplace-h44mk\" (UID: \"5264d4a2-bda1-4cdd-a63e-e924bd43bc82\") " pod="openshift-marketplace/redhat-marketplace-h44mk" Mar 17 10:10:08 crc kubenswrapper[4813]: I0317 10:10:08.117934 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fpr48\" (UniqueName: \"kubernetes.io/projected/5264d4a2-bda1-4cdd-a63e-e924bd43bc82-kube-api-access-fpr48\") pod \"redhat-marketplace-h44mk\" (UID: \"5264d4a2-bda1-4cdd-a63e-e924bd43bc82\") " pod="openshift-marketplace/redhat-marketplace-h44mk" Mar 17 10:10:08 crc kubenswrapper[4813]: I0317 10:10:08.117984 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5264d4a2-bda1-4cdd-a63e-e924bd43bc82-utilities\") pod \"redhat-marketplace-h44mk\" (UID: \"5264d4a2-bda1-4cdd-a63e-e924bd43bc82\") " pod="openshift-marketplace/redhat-marketplace-h44mk" Mar 17 10:10:08 crc kubenswrapper[4813]: I0317 10:10:08.118418 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5264d4a2-bda1-4cdd-a63e-e924bd43bc82-utilities\") pod \"redhat-marketplace-h44mk\" (UID: \"5264d4a2-bda1-4cdd-a63e-e924bd43bc82\") " pod="openshift-marketplace/redhat-marketplace-h44mk" Mar 17 10:10:08 crc kubenswrapper[4813]: I0317 10:10:08.118499 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5264d4a2-bda1-4cdd-a63e-e924bd43bc82-catalog-content\") pod \"redhat-marketplace-h44mk\" (UID: \"5264d4a2-bda1-4cdd-a63e-e924bd43bc82\") " pod="openshift-marketplace/redhat-marketplace-h44mk" Mar 17 10:10:08 crc kubenswrapper[4813]: I0317 10:10:08.138981 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fpr48\" (UniqueName: \"kubernetes.io/projected/5264d4a2-bda1-4cdd-a63e-e924bd43bc82-kube-api-access-fpr48\") pod \"redhat-marketplace-h44mk\" (UID: \"5264d4a2-bda1-4cdd-a63e-e924bd43bc82\") " pod="openshift-marketplace/redhat-marketplace-h44mk" Mar 17 10:10:08 crc kubenswrapper[4813]: I0317 10:10:08.284166 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h44mk" Mar 17 10:10:08 crc kubenswrapper[4813]: I0317 10:10:08.773328 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-h44mk"] Mar 17 10:10:09 crc kubenswrapper[4813]: I0317 10:10:09.330515 4813 generic.go:334] "Generic (PLEG): container finished" podID="5264d4a2-bda1-4cdd-a63e-e924bd43bc82" containerID="ed4cb8d317c0b54d3bc5bbaeace21558f499b7c32b88145afe748bf36efdd541" exitCode=0 Mar 17 10:10:09 crc kubenswrapper[4813]: I0317 10:10:09.330692 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h44mk" event={"ID":"5264d4a2-bda1-4cdd-a63e-e924bd43bc82","Type":"ContainerDied","Data":"ed4cb8d317c0b54d3bc5bbaeace21558f499b7c32b88145afe748bf36efdd541"} Mar 17 10:10:09 crc kubenswrapper[4813]: I0317 10:10:09.330911 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h44mk" event={"ID":"5264d4a2-bda1-4cdd-a63e-e924bd43bc82","Type":"ContainerStarted","Data":"933a35d7d7507d2aa182c7d5e36da028c51810fca46e0ce78e36ba42fc59c402"} Mar 17 10:10:10 crc kubenswrapper[4813]: I0317 10:10:10.348181 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h44mk" event={"ID":"5264d4a2-bda1-4cdd-a63e-e924bd43bc82","Type":"ContainerStarted","Data":"35145964c8cb6a07e9006faecd84bbca7ae2e194bf17353b5b3b743ff5040abc"} Mar 17 10:10:12 crc kubenswrapper[4813]: I0317 10:10:12.368511 4813 generic.go:334] "Generic (PLEG): container finished" podID="5264d4a2-bda1-4cdd-a63e-e924bd43bc82" containerID="35145964c8cb6a07e9006faecd84bbca7ae2e194bf17353b5b3b743ff5040abc" exitCode=0 Mar 17 10:10:12 crc kubenswrapper[4813]: I0317 10:10:12.368580 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h44mk" event={"ID":"5264d4a2-bda1-4cdd-a63e-e924bd43bc82","Type":"ContainerDied","Data":"35145964c8cb6a07e9006faecd84bbca7ae2e194bf17353b5b3b743ff5040abc"} Mar 17 10:10:13 crc kubenswrapper[4813]: I0317 10:10:13.383173 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h44mk" event={"ID":"5264d4a2-bda1-4cdd-a63e-e924bd43bc82","Type":"ContainerStarted","Data":"667e54b83910296ab2ca1545a9d2a2fde4d83c9d89c92c098671fc4cc2761bfd"} Mar 17 10:10:13 crc kubenswrapper[4813]: I0317 10:10:13.407635 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-h44mk" podStartSLOduration=2.895688615 podStartE2EDuration="6.407613837s" podCreationTimestamp="2026-03-17 10:10:07 +0000 UTC" firstStartedPulling="2026-03-17 10:10:09.333574677 +0000 UTC m=+3631.434378226" lastFinishedPulling="2026-03-17 10:10:12.845499949 +0000 UTC m=+3634.946303448" observedRunningTime="2026-03-17 10:10:13.404046416 +0000 UTC m=+3635.504849925" watchObservedRunningTime="2026-03-17 10:10:13.407613837 +0000 UTC m=+3635.508417356" Mar 17 10:10:14 crc kubenswrapper[4813]: I0317 10:10:14.114563 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 10:10:14 crc kubenswrapper[4813]: I0317 10:10:14.114649 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 10:10:16 crc kubenswrapper[4813]: I0317 10:10:16.417629 4813 generic.go:334] "Generic (PLEG): container finished" podID="68bcbff5-1452-47bf-9688-567614dd41f0" containerID="07714735dd2134ad863dcb2368769844bf3220bfb769ca505a47f92572e07e2f" exitCode=0 Mar 17 10:10:16 crc kubenswrapper[4813]: I0317 10:10:16.417683 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d49m5/crc-debug-mmw8t" event={"ID":"68bcbff5-1452-47bf-9688-567614dd41f0","Type":"ContainerDied","Data":"07714735dd2134ad863dcb2368769844bf3220bfb769ca505a47f92572e07e2f"} Mar 17 10:10:17 crc kubenswrapper[4813]: I0317 10:10:17.525845 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d49m5/crc-debug-mmw8t" Mar 17 10:10:17 crc kubenswrapper[4813]: I0317 10:10:17.559305 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-d49m5/crc-debug-mmw8t"] Mar 17 10:10:17 crc kubenswrapper[4813]: I0317 10:10:17.567591 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-d49m5/crc-debug-mmw8t"] Mar 17 10:10:17 crc kubenswrapper[4813]: I0317 10:10:17.614343 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kts2\" (UniqueName: \"kubernetes.io/projected/68bcbff5-1452-47bf-9688-567614dd41f0-kube-api-access-9kts2\") pod \"68bcbff5-1452-47bf-9688-567614dd41f0\" (UID: \"68bcbff5-1452-47bf-9688-567614dd41f0\") " Mar 17 10:10:17 crc kubenswrapper[4813]: I0317 10:10:17.615077 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/68bcbff5-1452-47bf-9688-567614dd41f0-host\") pod \"68bcbff5-1452-47bf-9688-567614dd41f0\" (UID: \"68bcbff5-1452-47bf-9688-567614dd41f0\") " Mar 17 10:10:17 crc kubenswrapper[4813]: I0317 10:10:17.615203 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/68bcbff5-1452-47bf-9688-567614dd41f0-host" (OuterVolumeSpecName: "host") pod "68bcbff5-1452-47bf-9688-567614dd41f0" (UID: "68bcbff5-1452-47bf-9688-567614dd41f0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 10:10:17 crc kubenswrapper[4813]: I0317 10:10:17.615716 4813 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/68bcbff5-1452-47bf-9688-567614dd41f0-host\") on node \"crc\" DevicePath \"\"" Mar 17 10:10:17 crc kubenswrapper[4813]: I0317 10:10:17.619193 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68bcbff5-1452-47bf-9688-567614dd41f0-kube-api-access-9kts2" (OuterVolumeSpecName: "kube-api-access-9kts2") pod "68bcbff5-1452-47bf-9688-567614dd41f0" (UID: "68bcbff5-1452-47bf-9688-567614dd41f0"). InnerVolumeSpecName "kube-api-access-9kts2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:10:17 crc kubenswrapper[4813]: I0317 10:10:17.717850 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kts2\" (UniqueName: \"kubernetes.io/projected/68bcbff5-1452-47bf-9688-567614dd41f0-kube-api-access-9kts2\") on node \"crc\" DevicePath \"\"" Mar 17 10:10:18 crc kubenswrapper[4813]: I0317 10:10:18.284695 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-h44mk" Mar 17 10:10:18 crc kubenswrapper[4813]: I0317 10:10:18.284983 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-h44mk" Mar 17 10:10:18 crc kubenswrapper[4813]: I0317 10:10:18.342300 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-h44mk" Mar 17 10:10:18 crc kubenswrapper[4813]: I0317 10:10:18.438959 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="affe9376880c74d7a35c421ee88d30f415334fc2b537d5ebafe005ffc0270514" Mar 17 10:10:18 crc kubenswrapper[4813]: I0317 10:10:18.439004 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d49m5/crc-debug-mmw8t" Mar 17 10:10:18 crc kubenswrapper[4813]: I0317 10:10:18.497216 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-h44mk" Mar 17 10:10:18 crc kubenswrapper[4813]: I0317 10:10:18.603781 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h44mk"] Mar 17 10:10:18 crc kubenswrapper[4813]: I0317 10:10:18.749082 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68bcbff5-1452-47bf-9688-567614dd41f0" path="/var/lib/kubelet/pods/68bcbff5-1452-47bf-9688-567614dd41f0/volumes" Mar 17 10:10:18 crc kubenswrapper[4813]: I0317 10:10:18.822552 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-d49m5/crc-debug-jmqbd"] Mar 17 10:10:18 crc kubenswrapper[4813]: E0317 10:10:18.823162 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68bcbff5-1452-47bf-9688-567614dd41f0" containerName="container-00" Mar 17 10:10:18 crc kubenswrapper[4813]: I0317 10:10:18.823200 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="68bcbff5-1452-47bf-9688-567614dd41f0" containerName="container-00" Mar 17 10:10:18 crc kubenswrapper[4813]: I0317 10:10:18.823622 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="68bcbff5-1452-47bf-9688-567614dd41f0" containerName="container-00" Mar 17 10:10:18 crc kubenswrapper[4813]: I0317 10:10:18.824641 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d49m5/crc-debug-jmqbd" Mar 17 10:10:18 crc kubenswrapper[4813]: I0317 10:10:18.970744 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8drzf\" (UniqueName: \"kubernetes.io/projected/9fdf8507-f8db-463e-9c9c-09cd26f58a46-kube-api-access-8drzf\") pod \"crc-debug-jmqbd\" (UID: \"9fdf8507-f8db-463e-9c9c-09cd26f58a46\") " pod="openshift-must-gather-d49m5/crc-debug-jmqbd" Mar 17 10:10:18 crc kubenswrapper[4813]: I0317 10:10:18.971186 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fdf8507-f8db-463e-9c9c-09cd26f58a46-host\") pod \"crc-debug-jmqbd\" (UID: \"9fdf8507-f8db-463e-9c9c-09cd26f58a46\") " pod="openshift-must-gather-d49m5/crc-debug-jmqbd" Mar 17 10:10:19 crc kubenswrapper[4813]: I0317 10:10:19.072921 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fdf8507-f8db-463e-9c9c-09cd26f58a46-host\") pod \"crc-debug-jmqbd\" (UID: \"9fdf8507-f8db-463e-9c9c-09cd26f58a46\") " pod="openshift-must-gather-d49m5/crc-debug-jmqbd" Mar 17 10:10:19 crc kubenswrapper[4813]: I0317 10:10:19.072998 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8drzf\" (UniqueName: \"kubernetes.io/projected/9fdf8507-f8db-463e-9c9c-09cd26f58a46-kube-api-access-8drzf\") pod \"crc-debug-jmqbd\" (UID: \"9fdf8507-f8db-463e-9c9c-09cd26f58a46\") " pod="openshift-must-gather-d49m5/crc-debug-jmqbd" Mar 17 10:10:19 crc kubenswrapper[4813]: I0317 10:10:19.073044 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fdf8507-f8db-463e-9c9c-09cd26f58a46-host\") pod \"crc-debug-jmqbd\" (UID: \"9fdf8507-f8db-463e-9c9c-09cd26f58a46\") " pod="openshift-must-gather-d49m5/crc-debug-jmqbd" Mar 17 10:10:19 crc kubenswrapper[4813]: I0317 10:10:19.090876 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8drzf\" (UniqueName: \"kubernetes.io/projected/9fdf8507-f8db-463e-9c9c-09cd26f58a46-kube-api-access-8drzf\") pod \"crc-debug-jmqbd\" (UID: \"9fdf8507-f8db-463e-9c9c-09cd26f58a46\") " pod="openshift-must-gather-d49m5/crc-debug-jmqbd" Mar 17 10:10:19 crc kubenswrapper[4813]: I0317 10:10:19.142855 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d49m5/crc-debug-jmqbd" Mar 17 10:10:19 crc kubenswrapper[4813]: I0317 10:10:19.452722 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d49m5/crc-debug-jmqbd" event={"ID":"9fdf8507-f8db-463e-9c9c-09cd26f58a46","Type":"ContainerStarted","Data":"2029cda7ecc42328ab03599a86967a8c885a5bfeddfb90555b49907725994eb8"} Mar 17 10:10:19 crc kubenswrapper[4813]: I0317 10:10:19.453055 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d49m5/crc-debug-jmqbd" event={"ID":"9fdf8507-f8db-463e-9c9c-09cd26f58a46","Type":"ContainerStarted","Data":"603509926ad40464a2327e0a07eb24f675502bb1dc0619d1088e374d284d2ca0"} Mar 17 10:10:19 crc kubenswrapper[4813]: I0317 10:10:19.475590 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-d49m5/crc-debug-jmqbd" podStartSLOduration=1.475566439 podStartE2EDuration="1.475566439s" podCreationTimestamp="2026-03-17 10:10:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 10:10:19.469147129 +0000 UTC m=+3641.569950668" watchObservedRunningTime="2026-03-17 10:10:19.475566439 +0000 UTC m=+3641.576369948" Mar 17 10:10:20 crc kubenswrapper[4813]: I0317 10:10:20.465068 4813 generic.go:334] "Generic (PLEG): container finished" podID="9fdf8507-f8db-463e-9c9c-09cd26f58a46" containerID="2029cda7ecc42328ab03599a86967a8c885a5bfeddfb90555b49907725994eb8" exitCode=0 Mar 17 10:10:20 crc kubenswrapper[4813]: I0317 10:10:20.465495 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-h44mk" podUID="5264d4a2-bda1-4cdd-a63e-e924bd43bc82" containerName="registry-server" containerID="cri-o://667e54b83910296ab2ca1545a9d2a2fde4d83c9d89c92c098671fc4cc2761bfd" gracePeriod=2 Mar 17 10:10:20 crc kubenswrapper[4813]: I0317 10:10:20.466036 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d49m5/crc-debug-jmqbd" event={"ID":"9fdf8507-f8db-463e-9c9c-09cd26f58a46","Type":"ContainerDied","Data":"2029cda7ecc42328ab03599a86967a8c885a5bfeddfb90555b49907725994eb8"} Mar 17 10:10:20 crc kubenswrapper[4813]: I0317 10:10:20.976427 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h44mk" Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.110630 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5264d4a2-bda1-4cdd-a63e-e924bd43bc82-utilities\") pod \"5264d4a2-bda1-4cdd-a63e-e924bd43bc82\" (UID: \"5264d4a2-bda1-4cdd-a63e-e924bd43bc82\") " Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.110736 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fpr48\" (UniqueName: \"kubernetes.io/projected/5264d4a2-bda1-4cdd-a63e-e924bd43bc82-kube-api-access-fpr48\") pod \"5264d4a2-bda1-4cdd-a63e-e924bd43bc82\" (UID: \"5264d4a2-bda1-4cdd-a63e-e924bd43bc82\") " Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.110842 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5264d4a2-bda1-4cdd-a63e-e924bd43bc82-catalog-content\") pod \"5264d4a2-bda1-4cdd-a63e-e924bd43bc82\" (UID: \"5264d4a2-bda1-4cdd-a63e-e924bd43bc82\") " Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.112067 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5264d4a2-bda1-4cdd-a63e-e924bd43bc82-utilities" (OuterVolumeSpecName: "utilities") pod "5264d4a2-bda1-4cdd-a63e-e924bd43bc82" (UID: "5264d4a2-bda1-4cdd-a63e-e924bd43bc82"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.125378 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5264d4a2-bda1-4cdd-a63e-e924bd43bc82-kube-api-access-fpr48" (OuterVolumeSpecName: "kube-api-access-fpr48") pod "5264d4a2-bda1-4cdd-a63e-e924bd43bc82" (UID: "5264d4a2-bda1-4cdd-a63e-e924bd43bc82"). InnerVolumeSpecName "kube-api-access-fpr48". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.136976 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5264d4a2-bda1-4cdd-a63e-e924bd43bc82-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5264d4a2-bda1-4cdd-a63e-e924bd43bc82" (UID: "5264d4a2-bda1-4cdd-a63e-e924bd43bc82"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.213373 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5264d4a2-bda1-4cdd-a63e-e924bd43bc82-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.213428 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fpr48\" (UniqueName: \"kubernetes.io/projected/5264d4a2-bda1-4cdd-a63e-e924bd43bc82-kube-api-access-fpr48\") on node \"crc\" DevicePath \"\"" Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.213447 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5264d4a2-bda1-4cdd-a63e-e924bd43bc82-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.479178 4813 generic.go:334] "Generic (PLEG): container finished" podID="5264d4a2-bda1-4cdd-a63e-e924bd43bc82" containerID="667e54b83910296ab2ca1545a9d2a2fde4d83c9d89c92c098671fc4cc2761bfd" exitCode=0 Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.479300 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-h44mk" Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.479285 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h44mk" event={"ID":"5264d4a2-bda1-4cdd-a63e-e924bd43bc82","Type":"ContainerDied","Data":"667e54b83910296ab2ca1545a9d2a2fde4d83c9d89c92c098671fc4cc2761bfd"} Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.480943 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-h44mk" event={"ID":"5264d4a2-bda1-4cdd-a63e-e924bd43bc82","Type":"ContainerDied","Data":"933a35d7d7507d2aa182c7d5e36da028c51810fca46e0ce78e36ba42fc59c402"} Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.481044 4813 scope.go:117] "RemoveContainer" containerID="667e54b83910296ab2ca1545a9d2a2fde4d83c9d89c92c098671fc4cc2761bfd" Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.561829 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d49m5/crc-debug-jmqbd" Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.562393 4813 scope.go:117] "RemoveContainer" containerID="35145964c8cb6a07e9006faecd84bbca7ae2e194bf17353b5b3b743ff5040abc" Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.592664 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-h44mk"] Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.605253 4813 scope.go:117] "RemoveContainer" containerID="ed4cb8d317c0b54d3bc5bbaeace21558f499b7c32b88145afe748bf36efdd541" Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.611184 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-h44mk"] Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.620097 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8drzf\" (UniqueName: \"kubernetes.io/projected/9fdf8507-f8db-463e-9c9c-09cd26f58a46-kube-api-access-8drzf\") pod \"9fdf8507-f8db-463e-9c9c-09cd26f58a46\" (UID: \"9fdf8507-f8db-463e-9c9c-09cd26f58a46\") " Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.620328 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fdf8507-f8db-463e-9c9c-09cd26f58a46-host\") pod \"9fdf8507-f8db-463e-9c9c-09cd26f58a46\" (UID: \"9fdf8507-f8db-463e-9c9c-09cd26f58a46\") " Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.620395 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9fdf8507-f8db-463e-9c9c-09cd26f58a46-host" (OuterVolumeSpecName: "host") pod "9fdf8507-f8db-463e-9c9c-09cd26f58a46" (UID: "9fdf8507-f8db-463e-9c9c-09cd26f58a46"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.620964 4813 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9fdf8507-f8db-463e-9c9c-09cd26f58a46-host\") on node \"crc\" DevicePath \"\"" Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.623726 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9fdf8507-f8db-463e-9c9c-09cd26f58a46-kube-api-access-8drzf" (OuterVolumeSpecName: "kube-api-access-8drzf") pod "9fdf8507-f8db-463e-9c9c-09cd26f58a46" (UID: "9fdf8507-f8db-463e-9c9c-09cd26f58a46"). InnerVolumeSpecName "kube-api-access-8drzf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.634828 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-d49m5/crc-debug-jmqbd"] Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.644725 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-d49m5/crc-debug-jmqbd"] Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.680470 4813 scope.go:117] "RemoveContainer" containerID="667e54b83910296ab2ca1545a9d2a2fde4d83c9d89c92c098671fc4cc2761bfd" Mar 17 10:10:21 crc kubenswrapper[4813]: E0317 10:10:21.681161 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"667e54b83910296ab2ca1545a9d2a2fde4d83c9d89c92c098671fc4cc2761bfd\": container with ID starting with 667e54b83910296ab2ca1545a9d2a2fde4d83c9d89c92c098671fc4cc2761bfd not found: ID does not exist" containerID="667e54b83910296ab2ca1545a9d2a2fde4d83c9d89c92c098671fc4cc2761bfd" Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.681199 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"667e54b83910296ab2ca1545a9d2a2fde4d83c9d89c92c098671fc4cc2761bfd"} err="failed to get container status \"667e54b83910296ab2ca1545a9d2a2fde4d83c9d89c92c098671fc4cc2761bfd\": rpc error: code = NotFound desc = could not find container \"667e54b83910296ab2ca1545a9d2a2fde4d83c9d89c92c098671fc4cc2761bfd\": container with ID starting with 667e54b83910296ab2ca1545a9d2a2fde4d83c9d89c92c098671fc4cc2761bfd not found: ID does not exist" Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.681223 4813 scope.go:117] "RemoveContainer" containerID="35145964c8cb6a07e9006faecd84bbca7ae2e194bf17353b5b3b743ff5040abc" Mar 17 10:10:21 crc kubenswrapper[4813]: E0317 10:10:21.681522 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35145964c8cb6a07e9006faecd84bbca7ae2e194bf17353b5b3b743ff5040abc\": container with ID starting with 35145964c8cb6a07e9006faecd84bbca7ae2e194bf17353b5b3b743ff5040abc not found: ID does not exist" containerID="35145964c8cb6a07e9006faecd84bbca7ae2e194bf17353b5b3b743ff5040abc" Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.681569 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35145964c8cb6a07e9006faecd84bbca7ae2e194bf17353b5b3b743ff5040abc"} err="failed to get container status \"35145964c8cb6a07e9006faecd84bbca7ae2e194bf17353b5b3b743ff5040abc\": rpc error: code = NotFound desc = could not find container \"35145964c8cb6a07e9006faecd84bbca7ae2e194bf17353b5b3b743ff5040abc\": container with ID starting with 35145964c8cb6a07e9006faecd84bbca7ae2e194bf17353b5b3b743ff5040abc not found: ID does not exist" Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.681622 4813 scope.go:117] "RemoveContainer" containerID="ed4cb8d317c0b54d3bc5bbaeace21558f499b7c32b88145afe748bf36efdd541" Mar 17 10:10:21 crc kubenswrapper[4813]: E0317 10:10:21.681933 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed4cb8d317c0b54d3bc5bbaeace21558f499b7c32b88145afe748bf36efdd541\": container with ID starting with ed4cb8d317c0b54d3bc5bbaeace21558f499b7c32b88145afe748bf36efdd541 not found: ID does not exist" containerID="ed4cb8d317c0b54d3bc5bbaeace21558f499b7c32b88145afe748bf36efdd541" Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.681965 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed4cb8d317c0b54d3bc5bbaeace21558f499b7c32b88145afe748bf36efdd541"} err="failed to get container status \"ed4cb8d317c0b54d3bc5bbaeace21558f499b7c32b88145afe748bf36efdd541\": rpc error: code = NotFound desc = could not find container \"ed4cb8d317c0b54d3bc5bbaeace21558f499b7c32b88145afe748bf36efdd541\": container with ID starting with ed4cb8d317c0b54d3bc5bbaeace21558f499b7c32b88145afe748bf36efdd541 not found: ID does not exist" Mar 17 10:10:21 crc kubenswrapper[4813]: I0317 10:10:21.733614 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8drzf\" (UniqueName: \"kubernetes.io/projected/9fdf8507-f8db-463e-9c9c-09cd26f58a46-kube-api-access-8drzf\") on node \"crc\" DevicePath \"\"" Mar 17 10:10:22 crc kubenswrapper[4813]: I0317 10:10:22.492120 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="603509926ad40464a2327e0a07eb24f675502bb1dc0619d1088e374d284d2ca0" Mar 17 10:10:22 crc kubenswrapper[4813]: I0317 10:10:22.492236 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d49m5/crc-debug-jmqbd" Mar 17 10:10:22 crc kubenswrapper[4813]: I0317 10:10:22.744959 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5264d4a2-bda1-4cdd-a63e-e924bd43bc82" path="/var/lib/kubelet/pods/5264d4a2-bda1-4cdd-a63e-e924bd43bc82/volumes" Mar 17 10:10:22 crc kubenswrapper[4813]: I0317 10:10:22.745825 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9fdf8507-f8db-463e-9c9c-09cd26f58a46" path="/var/lib/kubelet/pods/9fdf8507-f8db-463e-9c9c-09cd26f58a46/volumes" Mar 17 10:10:22 crc kubenswrapper[4813]: I0317 10:10:22.867511 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-d49m5/crc-debug-ts8pm"] Mar 17 10:10:22 crc kubenswrapper[4813]: E0317 10:10:22.868455 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5264d4a2-bda1-4cdd-a63e-e924bd43bc82" containerName="registry-server" Mar 17 10:10:22 crc kubenswrapper[4813]: I0317 10:10:22.868489 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5264d4a2-bda1-4cdd-a63e-e924bd43bc82" containerName="registry-server" Mar 17 10:10:22 crc kubenswrapper[4813]: E0317 10:10:22.868529 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5264d4a2-bda1-4cdd-a63e-e924bd43bc82" containerName="extract-utilities" Mar 17 10:10:22 crc kubenswrapper[4813]: I0317 10:10:22.868544 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5264d4a2-bda1-4cdd-a63e-e924bd43bc82" containerName="extract-utilities" Mar 17 10:10:22 crc kubenswrapper[4813]: E0317 10:10:22.868582 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9fdf8507-f8db-463e-9c9c-09cd26f58a46" containerName="container-00" Mar 17 10:10:22 crc kubenswrapper[4813]: I0317 10:10:22.868620 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9fdf8507-f8db-463e-9c9c-09cd26f58a46" containerName="container-00" Mar 17 10:10:22 crc kubenswrapper[4813]: E0317 10:10:22.868641 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5264d4a2-bda1-4cdd-a63e-e924bd43bc82" containerName="extract-content" Mar 17 10:10:22 crc kubenswrapper[4813]: I0317 10:10:22.868655 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5264d4a2-bda1-4cdd-a63e-e924bd43bc82" containerName="extract-content" Mar 17 10:10:22 crc kubenswrapper[4813]: I0317 10:10:22.868978 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9fdf8507-f8db-463e-9c9c-09cd26f58a46" containerName="container-00" Mar 17 10:10:22 crc kubenswrapper[4813]: I0317 10:10:22.869036 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5264d4a2-bda1-4cdd-a63e-e924bd43bc82" containerName="registry-server" Mar 17 10:10:22 crc kubenswrapper[4813]: I0317 10:10:22.870014 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d49m5/crc-debug-ts8pm" Mar 17 10:10:22 crc kubenswrapper[4813]: I0317 10:10:22.962044 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4-host\") pod \"crc-debug-ts8pm\" (UID: \"d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4\") " pod="openshift-must-gather-d49m5/crc-debug-ts8pm" Mar 17 10:10:22 crc kubenswrapper[4813]: I0317 10:10:22.962140 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-llc86\" (UniqueName: \"kubernetes.io/projected/d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4-kube-api-access-llc86\") pod \"crc-debug-ts8pm\" (UID: \"d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4\") " pod="openshift-must-gather-d49m5/crc-debug-ts8pm" Mar 17 10:10:23 crc kubenswrapper[4813]: I0317 10:10:23.064118 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4-host\") pod \"crc-debug-ts8pm\" (UID: \"d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4\") " pod="openshift-must-gather-d49m5/crc-debug-ts8pm" Mar 17 10:10:23 crc kubenswrapper[4813]: I0317 10:10:23.064239 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-llc86\" (UniqueName: \"kubernetes.io/projected/d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4-kube-api-access-llc86\") pod \"crc-debug-ts8pm\" (UID: \"d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4\") " pod="openshift-must-gather-d49m5/crc-debug-ts8pm" Mar 17 10:10:23 crc kubenswrapper[4813]: I0317 10:10:23.064339 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4-host\") pod \"crc-debug-ts8pm\" (UID: \"d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4\") " pod="openshift-must-gather-d49m5/crc-debug-ts8pm" Mar 17 10:10:23 crc kubenswrapper[4813]: I0317 10:10:23.104804 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-llc86\" (UniqueName: \"kubernetes.io/projected/d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4-kube-api-access-llc86\") pod \"crc-debug-ts8pm\" (UID: \"d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4\") " pod="openshift-must-gather-d49m5/crc-debug-ts8pm" Mar 17 10:10:23 crc kubenswrapper[4813]: I0317 10:10:23.195784 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d49m5/crc-debug-ts8pm" Mar 17 10:10:23 crc kubenswrapper[4813]: W0317 10:10:23.254133 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd58a8fbe_13bc_45d3_8dfb_e7df1e493ff4.slice/crio-11bcbe288b63223fb7bc285e894711b43d0b74d57973e24b39ee1b22a7d8f077 WatchSource:0}: Error finding container 11bcbe288b63223fb7bc285e894711b43d0b74d57973e24b39ee1b22a7d8f077: Status 404 returned error can't find the container with id 11bcbe288b63223fb7bc285e894711b43d0b74d57973e24b39ee1b22a7d8f077 Mar 17 10:10:23 crc kubenswrapper[4813]: I0317 10:10:23.508231 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d49m5/crc-debug-ts8pm" event={"ID":"d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4","Type":"ContainerStarted","Data":"11bcbe288b63223fb7bc285e894711b43d0b74d57973e24b39ee1b22a7d8f077"} Mar 17 10:10:24 crc kubenswrapper[4813]: I0317 10:10:24.523542 4813 generic.go:334] "Generic (PLEG): container finished" podID="d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4" containerID="305970afb4b3cb35949c3ff67c3993b5ec344b5773007b0174708a7772747c11" exitCode=0 Mar 17 10:10:24 crc kubenswrapper[4813]: I0317 10:10:24.523652 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d49m5/crc-debug-ts8pm" event={"ID":"d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4","Type":"ContainerDied","Data":"305970afb4b3cb35949c3ff67c3993b5ec344b5773007b0174708a7772747c11"} Mar 17 10:10:24 crc kubenswrapper[4813]: I0317 10:10:24.596741 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-d49m5/crc-debug-ts8pm"] Mar 17 10:10:24 crc kubenswrapper[4813]: I0317 10:10:24.608330 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-d49m5/crc-debug-ts8pm"] Mar 17 10:10:25 crc kubenswrapper[4813]: I0317 10:10:25.640263 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d49m5/crc-debug-ts8pm" Mar 17 10:10:25 crc kubenswrapper[4813]: I0317 10:10:25.715353 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4-host\") pod \"d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4\" (UID: \"d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4\") " Mar 17 10:10:25 crc kubenswrapper[4813]: I0317 10:10:25.715626 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4-host" (OuterVolumeSpecName: "host") pod "d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4" (UID: "d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 10:10:25 crc kubenswrapper[4813]: I0317 10:10:25.715638 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-llc86\" (UniqueName: \"kubernetes.io/projected/d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4-kube-api-access-llc86\") pod \"d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4\" (UID: \"d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4\") " Mar 17 10:10:25 crc kubenswrapper[4813]: I0317 10:10:25.716363 4813 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4-host\") on node \"crc\" DevicePath \"\"" Mar 17 10:10:25 crc kubenswrapper[4813]: I0317 10:10:25.721272 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4-kube-api-access-llc86" (OuterVolumeSpecName: "kube-api-access-llc86") pod "d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4" (UID: "d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4"). InnerVolumeSpecName "kube-api-access-llc86". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:10:25 crc kubenswrapper[4813]: I0317 10:10:25.818686 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-llc86\" (UniqueName: \"kubernetes.io/projected/d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4-kube-api-access-llc86\") on node \"crc\" DevicePath \"\"" Mar 17 10:10:26 crc kubenswrapper[4813]: I0317 10:10:26.547953 4813 scope.go:117] "RemoveContainer" containerID="305970afb4b3cb35949c3ff67c3993b5ec344b5773007b0174708a7772747c11" Mar 17 10:10:26 crc kubenswrapper[4813]: I0317 10:10:26.547982 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d49m5/crc-debug-ts8pm" Mar 17 10:10:26 crc kubenswrapper[4813]: I0317 10:10:26.740097 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4" path="/var/lib/kubelet/pods/d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4/volumes" Mar 17 10:10:40 crc kubenswrapper[4813]: I0317 10:10:40.032907 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-79bdd687db-2rnkr_80441abc-02b7-4a55-b4b9-2c3c7adf8ed2/barbican-api/0.log" Mar 17 10:10:40 crc kubenswrapper[4813]: I0317 10:10:40.288584 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-79bdd687db-2rnkr_80441abc-02b7-4a55-b4b9-2c3c7adf8ed2/barbican-api-log/0.log" Mar 17 10:10:40 crc kubenswrapper[4813]: I0317 10:10:40.416143 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5ff5576dbb-hlqck_2d44bf48-4229-453c-925d-9653d547de9e/barbican-keystone-listener/0.log" Mar 17 10:10:40 crc kubenswrapper[4813]: I0317 10:10:40.490879 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5ff5576dbb-hlqck_2d44bf48-4229-453c-925d-9653d547de9e/barbican-keystone-listener-log/0.log" Mar 17 10:10:40 crc kubenswrapper[4813]: I0317 10:10:40.595678 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-76556857b7-95bv6_91f4bba0-7f6d-43cf-8887-00b081856e89/barbican-worker/0.log" Mar 17 10:10:40 crc kubenswrapper[4813]: I0317 10:10:40.699025 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-76556857b7-95bv6_91f4bba0-7f6d-43cf-8887-00b081856e89/barbican-worker-log/0.log" Mar 17 10:10:40 crc kubenswrapper[4813]: I0317 10:10:40.817124 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg_f9166d03-0282-47e6-a80b-5b03f4183f62/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:10:40 crc kubenswrapper[4813]: I0317 10:10:40.907417 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ea9cbc63-ff93-4342-ab4c-7ddf7bf78577/ceilometer-central-agent/0.log" Mar 17 10:10:40 crc kubenswrapper[4813]: I0317 10:10:40.964905 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ea9cbc63-ff93-4342-ab4c-7ddf7bf78577/ceilometer-notification-agent/0.log" Mar 17 10:10:40 crc kubenswrapper[4813]: I0317 10:10:40.997186 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ea9cbc63-ff93-4342-ab4c-7ddf7bf78577/proxy-httpd/0.log" Mar 17 10:10:41 crc kubenswrapper[4813]: I0317 10:10:41.016093 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ea9cbc63-ff93-4342-ab4c-7ddf7bf78577/sg-core/0.log" Mar 17 10:10:41 crc kubenswrapper[4813]: I0317 10:10:41.161272 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_27d30ca5-0f9f-4f22-83a6-66c9e0942930/cinder-api-log/0.log" Mar 17 10:10:41 crc kubenswrapper[4813]: I0317 10:10:41.208083 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_27d30ca5-0f9f-4f22-83a6-66c9e0942930/cinder-api/0.log" Mar 17 10:10:41 crc kubenswrapper[4813]: I0317 10:10:41.317457 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_e2cad367-9c0f-4556-9e4c-b6173aa1b161/cinder-scheduler/0.log" Mar 17 10:10:41 crc kubenswrapper[4813]: I0317 10:10:41.403417 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_e2cad367-9c0f-4556-9e4c-b6173aa1b161/probe/0.log" Mar 17 10:10:41 crc kubenswrapper[4813]: I0317 10:10:41.476954 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-4kzll_c7908d9c-b027-4b03-8879-37968eea2f28/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:10:41 crc kubenswrapper[4813]: I0317 10:10:41.601810 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5_0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:10:41 crc kubenswrapper[4813]: I0317 10:10:41.708419 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-drzd4_3cceef96-6dbd-4173-b363-3bc4fbf4f598/init/0.log" Mar 17 10:10:41 crc kubenswrapper[4813]: I0317 10:10:41.876023 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-drzd4_3cceef96-6dbd-4173-b363-3bc4fbf4f598/init/0.log" Mar 17 10:10:41 crc kubenswrapper[4813]: I0317 10:10:41.890167 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-drzd4_3cceef96-6dbd-4173-b363-3bc4fbf4f598/dnsmasq-dns/0.log" Mar 17 10:10:41 crc kubenswrapper[4813]: I0317 10:10:41.921957 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-46np8_36536921-03a1-45dd-93b3-1d06e9c3adca/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:10:42 crc kubenswrapper[4813]: I0317 10:10:42.073294 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f4b4ddd1-d517-43cc-ac5e-636026efff5d/glance-httpd/0.log" Mar 17 10:10:42 crc kubenswrapper[4813]: I0317 10:10:42.116940 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f4b4ddd1-d517-43cc-ac5e-636026efff5d/glance-log/0.log" Mar 17 10:10:42 crc kubenswrapper[4813]: I0317 10:10:42.260960 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7056da71-3a67-4258-8f12-9ab7b50a83ea/glance-httpd/0.log" Mar 17 10:10:42 crc kubenswrapper[4813]: I0317 10:10:42.300802 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7056da71-3a67-4258-8f12-9ab7b50a83ea/glance-log/0.log" Mar 17 10:10:42 crc kubenswrapper[4813]: I0317 10:10:42.462250 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5dc999f6f8-t94dl_bf22966e-516d-40c2-975c-c3e41122b8d2/horizon/0.log" Mar 17 10:10:42 crc kubenswrapper[4813]: I0317 10:10:42.576908 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt_d8cecca4-b584-4537-810f-02c1818d2e99/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:10:42 crc kubenswrapper[4813]: I0317 10:10:42.765512 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5dc999f6f8-t94dl_bf22966e-516d-40c2-975c-c3e41122b8d2/horizon-log/0.log" Mar 17 10:10:42 crc kubenswrapper[4813]: I0317 10:10:42.793637 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-64lbt_63ad7386-248c-4dc8-a217-fb2cf3d4ef82/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:10:43 crc kubenswrapper[4813]: I0317 10:10:43.009674 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29562361-p6pcq_8a4ef497-31ad-415a-8c23-1673c0e279cb/keystone-cron/0.log" Mar 17 10:10:43 crc kubenswrapper[4813]: I0317 10:10:43.035393 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7fb6796ddc-58j2l_fd0582a1-002f-452c-828a-406a4d945f28/keystone-api/0.log" Mar 17 10:10:43 crc kubenswrapper[4813]: I0317 10:10:43.166953 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_9e83f1d1-1126-49a5-9a24-1694e7616b61/kube-state-metrics/0.log" Mar 17 10:10:43 crc kubenswrapper[4813]: I0317 10:10:43.248721 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8_71e0074f-5ee2-4f59-9184-c23495021bfd/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:10:43 crc kubenswrapper[4813]: I0317 10:10:43.569133 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-76bd9cc6f7-x2p6w_7d7e90bc-8bc2-4c06-9131-405fa21670df/neutron-httpd/0.log" Mar 17 10:10:43 crc kubenswrapper[4813]: I0317 10:10:43.600316 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-76bd9cc6f7-x2p6w_7d7e90bc-8bc2-4c06-9131-405fa21670df/neutron-api/0.log" Mar 17 10:10:43 crc kubenswrapper[4813]: I0317 10:10:43.765006 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj_0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:10:44 crc kubenswrapper[4813]: I0317 10:10:44.113747 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 10:10:44 crc kubenswrapper[4813]: I0317 10:10:44.113794 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 10:10:44 crc kubenswrapper[4813]: I0317 10:10:44.306324 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_08fc5042-2349-4425-a338-e9f9ab5cc125/nova-cell0-conductor-conductor/0.log" Mar 17 10:10:44 crc kubenswrapper[4813]: I0317 10:10:44.362122 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_04ea2c77-25c5-45fc-a375-cf04b185381c/nova-api-log/0.log" Mar 17 10:10:44 crc kubenswrapper[4813]: I0317 10:10:44.519314 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_04ea2c77-25c5-45fc-a375-cf04b185381c/nova-api-api/0.log" Mar 17 10:10:44 crc kubenswrapper[4813]: I0317 10:10:44.647168 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_c8cfa16e-a939-4e80-9e09-2a632e7bd29e/nova-cell1-conductor-conductor/0.log" Mar 17 10:10:44 crc kubenswrapper[4813]: I0317 10:10:44.712749 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_9cc3a926-a367-45b1-9d38-6f673720e71e/nova-cell1-novncproxy-novncproxy/0.log" Mar 17 10:10:44 crc kubenswrapper[4813]: I0317 10:10:44.871044 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-6z5zk_bd00eea2-860c-40ac-8147-41992af388ec/nova-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:10:45 crc kubenswrapper[4813]: I0317 10:10:45.000530 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_82bd221d-6ac0-4eb2-b709-3de76c656745/nova-metadata-log/0.log" Mar 17 10:10:45 crc kubenswrapper[4813]: I0317 10:10:45.305048 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1/mysql-bootstrap/0.log" Mar 17 10:10:45 crc kubenswrapper[4813]: I0317 10:10:45.360404 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_aa41f41c-74ec-44a3-a913-afd1a44d4d04/nova-scheduler-scheduler/0.log" Mar 17 10:10:45 crc kubenswrapper[4813]: I0317 10:10:45.375159 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_82bd221d-6ac0-4eb2-b709-3de76c656745/nova-metadata-metadata/0.log" Mar 17 10:10:45 crc kubenswrapper[4813]: I0317 10:10:45.558693 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1/mysql-bootstrap/0.log" Mar 17 10:10:45 crc kubenswrapper[4813]: I0317 10:10:45.559167 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1/galera/0.log" Mar 17 10:10:45 crc kubenswrapper[4813]: I0317 10:10:45.574085 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9ad18f78-70c7-4b1d-bda7-ed338fdcdd64/mysql-bootstrap/0.log" Mar 17 10:10:45 crc kubenswrapper[4813]: I0317 10:10:45.847960 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9ad18f78-70c7-4b1d-bda7-ed338fdcdd64/mysql-bootstrap/0.log" Mar 17 10:10:45 crc kubenswrapper[4813]: I0317 10:10:45.852553 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_c532fc6b-8310-4322-b09c-bd21b19154ec/openstackclient/0.log" Mar 17 10:10:45 crc kubenswrapper[4813]: I0317 10:10:45.875035 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9ad18f78-70c7-4b1d-bda7-ed338fdcdd64/galera/0.log" Mar 17 10:10:46 crc kubenswrapper[4813]: I0317 10:10:46.064317 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-btlxg_8cbfe69c-a2c5-40bf-a8d2-e78755f060f2/openstack-network-exporter/0.log" Mar 17 10:10:46 crc kubenswrapper[4813]: I0317 10:10:46.107613 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bt47r_3752b62f-49dd-46a7-ae34-9b6b20343a09/ovsdb-server-init/0.log" Mar 17 10:10:46 crc kubenswrapper[4813]: I0317 10:10:46.292057 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bt47r_3752b62f-49dd-46a7-ae34-9b6b20343a09/ovsdb-server-init/0.log" Mar 17 10:10:46 crc kubenswrapper[4813]: I0317 10:10:46.329086 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bt47r_3752b62f-49dd-46a7-ae34-9b6b20343a09/ovsdb-server/0.log" Mar 17 10:10:46 crc kubenswrapper[4813]: I0317 10:10:46.352452 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bt47r_3752b62f-49dd-46a7-ae34-9b6b20343a09/ovs-vswitchd/0.log" Mar 17 10:10:46 crc kubenswrapper[4813]: I0317 10:10:46.471786 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-vdc6c_4090993b-51ce-4ce3-a6d6-a1501ab3ba05/ovn-controller/0.log" Mar 17 10:10:46 crc kubenswrapper[4813]: I0317 10:10:46.591110 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-c5ttr_5e4d8ae8-8ff6-4969-a349-813dde113094/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:10:46 crc kubenswrapper[4813]: I0317 10:10:46.679896 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_b433a82f-1948-4fdc-a395-86a0c07fee36/openstack-network-exporter/0.log" Mar 17 10:10:46 crc kubenswrapper[4813]: I0317 10:10:46.790645 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_b433a82f-1948-4fdc-a395-86a0c07fee36/ovn-northd/0.log" Mar 17 10:10:46 crc kubenswrapper[4813]: I0317 10:10:46.863882 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_34e286fb-7dd1-4e85-89e3-e926b232f5a6/ovsdbserver-nb/0.log" Mar 17 10:10:46 crc kubenswrapper[4813]: I0317 10:10:46.881979 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_34e286fb-7dd1-4e85-89e3-e926b232f5a6/openstack-network-exporter/0.log" Mar 17 10:10:46 crc kubenswrapper[4813]: I0317 10:10:46.980140 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7e4c8a33-829e-4aec-be7c-0dad92ce3916/openstack-network-exporter/0.log" Mar 17 10:10:47 crc kubenswrapper[4813]: I0317 10:10:47.084571 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7e4c8a33-829e-4aec-be7c-0dad92ce3916/ovsdbserver-sb/0.log" Mar 17 10:10:47 crc kubenswrapper[4813]: I0317 10:10:47.307918 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-557566c676-bbf8q_46eacf89-88ff-4be7-a8c1-e90784324da2/placement-log/0.log" Mar 17 10:10:47 crc kubenswrapper[4813]: I0317 10:10:47.331012 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-557566c676-bbf8q_46eacf89-88ff-4be7-a8c1-e90784324da2/placement-api/0.log" Mar 17 10:10:47 crc kubenswrapper[4813]: I0317 10:10:47.412518 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cfc379d0-d53a-4640-8d3f-a883db1aefa9/setup-container/0.log" Mar 17 10:10:47 crc kubenswrapper[4813]: I0317 10:10:47.532915 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cfc379d0-d53a-4640-8d3f-a883db1aefa9/setup-container/0.log" Mar 17 10:10:47 crc kubenswrapper[4813]: I0317 10:10:47.555232 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cfc379d0-d53a-4640-8d3f-a883db1aefa9/rabbitmq/0.log" Mar 17 10:10:47 crc kubenswrapper[4813]: I0317 10:10:47.624940 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b/setup-container/0.log" Mar 17 10:10:47 crc kubenswrapper[4813]: I0317 10:10:47.824531 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b/rabbitmq/0.log" Mar 17 10:10:47 crc kubenswrapper[4813]: I0317 10:10:47.853007 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk_60339fa3-6329-4aa7-a958-d28be7f562cc/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:10:47 crc kubenswrapper[4813]: I0317 10:10:47.867114 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b/setup-container/0.log" Mar 17 10:10:48 crc kubenswrapper[4813]: I0317 10:10:48.030080 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-5blh4_c0a603db-d69a-47a6-8228-862e4ad835ee/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:10:48 crc kubenswrapper[4813]: I0317 10:10:48.170617 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr_df99907b-2dfb-436f-b9a0-5c62dce82672/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:10:48 crc kubenswrapper[4813]: I0317 10:10:48.276109 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-jhl6m_540d96f7-a4a1-490f-9c57-997a2c6dc9ab/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:10:48 crc kubenswrapper[4813]: I0317 10:10:48.357807 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-ht9q9_1d6a2187-6711-4b3f-988a-845c5b7e18f7/ssh-known-hosts-edpm-deployment/0.log" Mar 17 10:10:48 crc kubenswrapper[4813]: I0317 10:10:48.583951 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-8557f6579f-shvsw_e79b20b0-682e-4d16-bb56-64f0c4ec0202/proxy-server/0.log" Mar 17 10:10:48 crc kubenswrapper[4813]: I0317 10:10:48.634541 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-8557f6579f-shvsw_e79b20b0-682e-4d16-bb56-64f0c4ec0202/proxy-httpd/0.log" Mar 17 10:10:48 crc kubenswrapper[4813]: I0317 10:10:48.699188 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-rrfrc_625a6e48-f069-49a9-b11a-342e12dffece/swift-ring-rebalance/0.log" Mar 17 10:10:48 crc kubenswrapper[4813]: I0317 10:10:48.818474 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/account-auditor/0.log" Mar 17 10:10:48 crc kubenswrapper[4813]: I0317 10:10:48.859403 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/account-reaper/0.log" Mar 17 10:10:49 crc kubenswrapper[4813]: I0317 10:10:49.059278 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/account-replicator/0.log" Mar 17 10:10:49 crc kubenswrapper[4813]: I0317 10:10:49.133867 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/account-server/0.log" Mar 17 10:10:49 crc kubenswrapper[4813]: I0317 10:10:49.183299 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/container-replicator/0.log" Mar 17 10:10:49 crc kubenswrapper[4813]: I0317 10:10:49.185821 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/container-auditor/0.log" Mar 17 10:10:49 crc kubenswrapper[4813]: I0317 10:10:49.243559 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/container-server/0.log" Mar 17 10:10:49 crc kubenswrapper[4813]: I0317 10:10:49.321588 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/container-updater/0.log" Mar 17 10:10:49 crc kubenswrapper[4813]: I0317 10:10:49.339177 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/object-expirer/0.log" Mar 17 10:10:49 crc kubenswrapper[4813]: I0317 10:10:49.459282 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/object-auditor/0.log" Mar 17 10:10:49 crc kubenswrapper[4813]: I0317 10:10:49.483942 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/object-server/0.log" Mar 17 10:10:49 crc kubenswrapper[4813]: I0317 10:10:49.485971 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/object-replicator/0.log" Mar 17 10:10:49 crc kubenswrapper[4813]: I0317 10:10:49.541948 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/object-updater/0.log" Mar 17 10:10:49 crc kubenswrapper[4813]: I0317 10:10:49.667490 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/rsync/0.log" Mar 17 10:10:49 crc kubenswrapper[4813]: I0317 10:10:49.671928 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/swift-recon-cron/0.log" Mar 17 10:10:49 crc kubenswrapper[4813]: I0317 10:10:49.882444 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-972kp_cd1a58de-c5c8-4b15-a894-47baac35d6e2/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:10:49 crc kubenswrapper[4813]: I0317 10:10:49.888541 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_005f04b0-01f5-4bed-9d2c-2269ab67d27f/tempest-tests-tempest-tests-runner/0.log" Mar 17 10:10:50 crc kubenswrapper[4813]: I0317 10:10:50.064017 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_5b58dbf4-d51d-4fe8-90d0-d295217084e1/test-operator-logs-container/0.log" Mar 17 10:10:50 crc kubenswrapper[4813]: I0317 10:10:50.149168 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm_0c2d5562-76da-433b-9840-59385eeb872c/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:10:50 crc kubenswrapper[4813]: I0317 10:10:50.749057 4813 scope.go:117] "RemoveContainer" containerID="82bb83955c3ed8ecebc47d512544710592de097b9756511eabf6f3c98f25f159" Mar 17 10:10:57 crc kubenswrapper[4813]: I0317 10:10:57.848729 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_c43912b9-949e-4d64-ae70-e59594cc329a/memcached/0.log" Mar 17 10:11:14 crc kubenswrapper[4813]: I0317 10:11:14.114221 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 10:11:14 crc kubenswrapper[4813]: I0317 10:11:14.114799 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 10:11:14 crc kubenswrapper[4813]: I0317 10:11:14.114860 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 10:11:14 crc kubenswrapper[4813]: I0317 10:11:14.115765 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9"} pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 10:11:14 crc kubenswrapper[4813]: I0317 10:11:14.115808 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" containerID="cri-o://bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" gracePeriod=600 Mar 17 10:11:14 crc kubenswrapper[4813]: E0317 10:11:14.233368 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:11:14 crc kubenswrapper[4813]: I0317 10:11:14.962221 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd_fa9ca0da-b53a-42bc-ace9-49579d21ead8/util/0.log" Mar 17 10:11:15 crc kubenswrapper[4813]: I0317 10:11:15.029643 4813 generic.go:334] "Generic (PLEG): container finished" podID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" exitCode=0 Mar 17 10:11:15 crc kubenswrapper[4813]: I0317 10:11:15.029685 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerDied","Data":"bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9"} Mar 17 10:11:15 crc kubenswrapper[4813]: I0317 10:11:15.029759 4813 scope.go:117] "RemoveContainer" containerID="586f3d43249f84824ce70b39efe9e7b1dbced1e1e5bd5f1cc753140f0d665cb9" Mar 17 10:11:15 crc kubenswrapper[4813]: I0317 10:11:15.030521 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:11:15 crc kubenswrapper[4813]: E0317 10:11:15.030885 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:11:15 crc kubenswrapper[4813]: I0317 10:11:15.176896 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd_fa9ca0da-b53a-42bc-ace9-49579d21ead8/pull/0.log" Mar 17 10:11:15 crc kubenswrapper[4813]: I0317 10:11:15.204167 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd_fa9ca0da-b53a-42bc-ace9-49579d21ead8/pull/0.log" Mar 17 10:11:15 crc kubenswrapper[4813]: I0317 10:11:15.213253 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd_fa9ca0da-b53a-42bc-ace9-49579d21ead8/util/0.log" Mar 17 10:11:15 crc kubenswrapper[4813]: I0317 10:11:15.368804 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd_fa9ca0da-b53a-42bc-ace9-49579d21ead8/pull/0.log" Mar 17 10:11:15 crc kubenswrapper[4813]: I0317 10:11:15.380765 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd_fa9ca0da-b53a-42bc-ace9-49579d21ead8/util/0.log" Mar 17 10:11:15 crc kubenswrapper[4813]: I0317 10:11:15.413469 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd_fa9ca0da-b53a-42bc-ace9-49579d21ead8/extract/0.log" Mar 17 10:11:15 crc kubenswrapper[4813]: I0317 10:11:15.585666 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-59bc569d95-xkfkm_227f29cf-ec7d-40e9-b28c-c1ccae9d880e/manager/0.log" Mar 17 10:11:15 crc kubenswrapper[4813]: I0317 10:11:15.755725 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-588d4d986b-tn4pv_9b3cf477-8c47-4a38-beee-8bd3f70977e0/manager/0.log" Mar 17 10:11:15 crc kubenswrapper[4813]: I0317 10:11:15.985332 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-79df6bcc97-wzmxx_3c5a1075-8105-401b-afb0-00014dafed0a/manager/0.log" Mar 17 10:11:16 crc kubenswrapper[4813]: I0317 10:11:16.061031 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-67dd5f86f5-hng2s_de6c9679-9f9f-481d-85d9-218c9199088e/manager/0.log" Mar 17 10:11:16 crc kubenswrapper[4813]: I0317 10:11:16.202630 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-8464cc45fb-tfhsc_668691cc-39f3-4b04-8ed6-bc86f9965f12/manager/0.log" Mar 17 10:11:16 crc kubenswrapper[4813]: I0317 10:11:16.458417 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6f787dddc9-4ds8c_80f14d5d-c8b9-4068-9469-ea8eff96e9e5/manager/0.log" Mar 17 10:11:16 crc kubenswrapper[4813]: I0317 10:11:16.546377 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-8d58dc466-9cqth_1b2b8088-31e5-49af-91b4-a3fbfe797af5/manager/0.log" Mar 17 10:11:16 crc kubenswrapper[4813]: I0317 10:11:16.701439 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7b9c774f96-7965b_c9ea4c59-d87b-4e58-ac8f-2786f8f32f02/manager/0.log" Mar 17 10:11:16 crc kubenswrapper[4813]: I0317 10:11:16.727633 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-768b96df4c-c966n_fd3396be-4825-4edc-a41f-443af1413e3b/manager/0.log" Mar 17 10:11:16 crc kubenswrapper[4813]: I0317 10:11:16.835759 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-55f864c847-dbs68_5e7b137d-9b58-448b-91a4-3a069dfb4d10/manager/0.log" Mar 17 10:11:16 crc kubenswrapper[4813]: I0317 10:11:16.897789 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67ccfc9778-ngd7n_0e3eba88-a4a4-4c6d-9e90-5dba41268494/manager/0.log" Mar 17 10:11:17 crc kubenswrapper[4813]: I0317 10:11:17.081534 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-767865f676-mnsb4_c3419efc-507c-46bc-aedb-ef5eb3f0f917/manager/0.log" Mar 17 10:11:17 crc kubenswrapper[4813]: I0317 10:11:17.184806 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5d488d59fb-tqg57_b4b9a61e-819f-4da5-a28a-8e49c0a3f7cd/manager/0.log" Mar 17 10:11:17 crc kubenswrapper[4813]: I0317 10:11:17.295715 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5b9f45d989-mps6t_c6e1a616-6d48-4ef1-93cb-62caed2de963/manager/0.log" Mar 17 10:11:17 crc kubenswrapper[4813]: I0317 10:11:17.391336 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-89d64c458-g5vgd_8e80dcf7-dae9-4faa-ab3f-b547b570926b/manager/0.log" Mar 17 10:11:17 crc kubenswrapper[4813]: I0317 10:11:17.748556 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-z8wcr_e56922bf-05f4-474c-b883-f1ca16e3e1cd/registry-server/0.log" Mar 17 10:11:17 crc kubenswrapper[4813]: I0317 10:11:17.783068 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-56ddf7fcb8-7hkdj_6dc9eebf-39dd-4030-a091-5626832fd141/operator/0.log" Mar 17 10:11:18 crc kubenswrapper[4813]: I0317 10:11:18.078363 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-884679f54-7d66d_76eb47fe-8f18-4f98-a8ba-d89169d0de93/manager/0.log" Mar 17 10:11:18 crc kubenswrapper[4813]: I0317 10:11:18.108778 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5784578c99-hcsvd_968158c3-2dfc-439e-9d20-8706d611fec1/manager/0.log" Mar 17 10:11:18 crc kubenswrapper[4813]: I0317 10:11:18.316428 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-rxzjs_bcd2954f-b240-4cc3-8387-15b0e64a9721/operator/0.log" Mar 17 10:11:18 crc kubenswrapper[4813]: I0317 10:11:18.438830 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-c674c5965-njxx7_ac052585-f733-47d5-91b9-87ad7957e511/manager/0.log" Mar 17 10:11:18 crc kubenswrapper[4813]: I0317 10:11:18.681012 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-d6b694c5-52xx2_ec48dbf8-730b-4a2e-9986-03d8f66de013/manager/0.log" Mar 17 10:11:18 crc kubenswrapper[4813]: I0317 10:11:18.811189 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-65f8859ffd-mxfzh_db193bcc-748f-4830-873c-48a2f9229b40/manager/0.log" Mar 17 10:11:18 crc kubenswrapper[4813]: I0317 10:11:18.855858 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-r2ksk_860ccadc-ca76-4523-9543-2b0960eb77a9/manager/0.log" Mar 17 10:11:18 crc kubenswrapper[4813]: I0317 10:11:18.939527 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6c4d75f7f9-wrtl9_520d3ab3-78d7-4be3-8466-9dadfce15c2c/manager/0.log" Mar 17 10:11:27 crc kubenswrapper[4813]: I0317 10:11:27.731891 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:11:27 crc kubenswrapper[4813]: E0317 10:11:27.732935 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:11:37 crc kubenswrapper[4813]: I0317 10:11:37.631443 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-j5djq_37f3773a-2367-411b-b8e4-c4ae96970de3/control-plane-machine-set-operator/0.log" Mar 17 10:11:37 crc kubenswrapper[4813]: I0317 10:11:37.805195 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-sk47b_feae3098-e93c-4dff-9088-a38db6a5d929/machine-api-operator/0.log" Mar 17 10:11:37 crc kubenswrapper[4813]: I0317 10:11:37.816250 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-sk47b_feae3098-e93c-4dff-9088-a38db6a5d929/kube-rbac-proxy/0.log" Mar 17 10:11:39 crc kubenswrapper[4813]: I0317 10:11:39.732139 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:11:39 crc kubenswrapper[4813]: E0317 10:11:39.732700 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:11:50 crc kubenswrapper[4813]: I0317 10:11:50.731430 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:11:50 crc kubenswrapper[4813]: E0317 10:11:50.732476 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:11:51 crc kubenswrapper[4813]: I0317 10:11:51.042390 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-knkd2_08c8f504-aa0d-47b9-ab86-a4b440c6f446/cert-manager-controller/0.log" Mar 17 10:11:51 crc kubenswrapper[4813]: I0317 10:11:51.170019 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-slrw5_7ff1e0fd-1a34-4f70-97c0-9fbb47f0b52b/cert-manager-cainjector/0.log" Mar 17 10:11:51 crc kubenswrapper[4813]: I0317 10:11:51.226146 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-hdhpn_d949287d-f3e6-4ac8-8c16-a0d0652ae302/cert-manager-webhook/0.log" Mar 17 10:12:00 crc kubenswrapper[4813]: I0317 10:12:00.153305 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562372-jrqr7"] Mar 17 10:12:00 crc kubenswrapper[4813]: E0317 10:12:00.154312 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4" containerName="container-00" Mar 17 10:12:00 crc kubenswrapper[4813]: I0317 10:12:00.154325 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4" containerName="container-00" Mar 17 10:12:00 crc kubenswrapper[4813]: I0317 10:12:00.154581 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="d58a8fbe-13bc-45d3-8dfb-e7df1e493ff4" containerName="container-00" Mar 17 10:12:00 crc kubenswrapper[4813]: I0317 10:12:00.155282 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562372-jrqr7" Mar 17 10:12:00 crc kubenswrapper[4813]: I0317 10:12:00.160808 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 10:12:00 crc kubenswrapper[4813]: I0317 10:12:00.161060 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 10:12:00 crc kubenswrapper[4813]: I0317 10:12:00.161267 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 10:12:00 crc kubenswrapper[4813]: I0317 10:12:00.176050 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562372-jrqr7"] Mar 17 10:12:00 crc kubenswrapper[4813]: I0317 10:12:00.250749 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw2jz\" (UniqueName: \"kubernetes.io/projected/b7b6e6f5-528c-40fa-8f42-892a57e71fee-kube-api-access-gw2jz\") pod \"auto-csr-approver-29562372-jrqr7\" (UID: \"b7b6e6f5-528c-40fa-8f42-892a57e71fee\") " pod="openshift-infra/auto-csr-approver-29562372-jrqr7" Mar 17 10:12:00 crc kubenswrapper[4813]: I0317 10:12:00.353310 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw2jz\" (UniqueName: \"kubernetes.io/projected/b7b6e6f5-528c-40fa-8f42-892a57e71fee-kube-api-access-gw2jz\") pod \"auto-csr-approver-29562372-jrqr7\" (UID: \"b7b6e6f5-528c-40fa-8f42-892a57e71fee\") " pod="openshift-infra/auto-csr-approver-29562372-jrqr7" Mar 17 10:12:00 crc kubenswrapper[4813]: I0317 10:12:00.371547 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw2jz\" (UniqueName: \"kubernetes.io/projected/b7b6e6f5-528c-40fa-8f42-892a57e71fee-kube-api-access-gw2jz\") pod \"auto-csr-approver-29562372-jrqr7\" (UID: \"b7b6e6f5-528c-40fa-8f42-892a57e71fee\") " pod="openshift-infra/auto-csr-approver-29562372-jrqr7" Mar 17 10:12:00 crc kubenswrapper[4813]: I0317 10:12:00.524324 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562372-jrqr7" Mar 17 10:12:00 crc kubenswrapper[4813]: I0317 10:12:00.951834 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562372-jrqr7"] Mar 17 10:12:01 crc kubenswrapper[4813]: I0317 10:12:01.474805 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562372-jrqr7" event={"ID":"b7b6e6f5-528c-40fa-8f42-892a57e71fee","Type":"ContainerStarted","Data":"eec39811a0873e08874c644be4af6ad6407357cf661b45c57b9124c4ad5da82b"} Mar 17 10:12:02 crc kubenswrapper[4813]: I0317 10:12:02.484695 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562372-jrqr7" event={"ID":"b7b6e6f5-528c-40fa-8f42-892a57e71fee","Type":"ContainerStarted","Data":"71f20f0035a4216c4aba6e5b93bc999c58c77eb58e565bc3704fca0d4432bccc"} Mar 17 10:12:02 crc kubenswrapper[4813]: I0317 10:12:02.515179 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562372-jrqr7" podStartSLOduration=1.351613747 podStartE2EDuration="2.51515914s" podCreationTimestamp="2026-03-17 10:12:00 +0000 UTC" firstStartedPulling="2026-03-17 10:12:00.968453506 +0000 UTC m=+3743.069257045" lastFinishedPulling="2026-03-17 10:12:02.131998909 +0000 UTC m=+3744.232802438" observedRunningTime="2026-03-17 10:12:02.500952017 +0000 UTC m=+3744.601755526" watchObservedRunningTime="2026-03-17 10:12:02.51515914 +0000 UTC m=+3744.615962639" Mar 17 10:12:02 crc kubenswrapper[4813]: I0317 10:12:02.730948 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:12:02 crc kubenswrapper[4813]: E0317 10:12:02.731359 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:12:03 crc kubenswrapper[4813]: I0317 10:12:03.500097 4813 generic.go:334] "Generic (PLEG): container finished" podID="b7b6e6f5-528c-40fa-8f42-892a57e71fee" containerID="71f20f0035a4216c4aba6e5b93bc999c58c77eb58e565bc3704fca0d4432bccc" exitCode=0 Mar 17 10:12:03 crc kubenswrapper[4813]: I0317 10:12:03.500152 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562372-jrqr7" event={"ID":"b7b6e6f5-528c-40fa-8f42-892a57e71fee","Type":"ContainerDied","Data":"71f20f0035a4216c4aba6e5b93bc999c58c77eb58e565bc3704fca0d4432bccc"} Mar 17 10:12:04 crc kubenswrapper[4813]: I0317 10:12:04.837081 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562372-jrqr7" Mar 17 10:12:04 crc kubenswrapper[4813]: I0317 10:12:04.941761 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gw2jz\" (UniqueName: \"kubernetes.io/projected/b7b6e6f5-528c-40fa-8f42-892a57e71fee-kube-api-access-gw2jz\") pod \"b7b6e6f5-528c-40fa-8f42-892a57e71fee\" (UID: \"b7b6e6f5-528c-40fa-8f42-892a57e71fee\") " Mar 17 10:12:04 crc kubenswrapper[4813]: I0317 10:12:04.947357 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7b6e6f5-528c-40fa-8f42-892a57e71fee-kube-api-access-gw2jz" (OuterVolumeSpecName: "kube-api-access-gw2jz") pod "b7b6e6f5-528c-40fa-8f42-892a57e71fee" (UID: "b7b6e6f5-528c-40fa-8f42-892a57e71fee"). InnerVolumeSpecName "kube-api-access-gw2jz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:12:05 crc kubenswrapper[4813]: I0317 10:12:05.043738 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gw2jz\" (UniqueName: \"kubernetes.io/projected/b7b6e6f5-528c-40fa-8f42-892a57e71fee-kube-api-access-gw2jz\") on node \"crc\" DevicePath \"\"" Mar 17 10:12:05 crc kubenswrapper[4813]: I0317 10:12:05.189101 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-4jmsm_89acc01c-091e-402c-a206-8fbc201f0884/nmstate-console-plugin/0.log" Mar 17 10:12:05 crc kubenswrapper[4813]: I0317 10:12:05.317061 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-zckpn_8813052b-cc40-4fe7-b4de-7aeb3832d4bb/nmstate-handler/0.log" Mar 17 10:12:05 crc kubenswrapper[4813]: I0317 10:12:05.368053 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-k2x5n_0d0f7c9a-ffef-4d0b-ae31-0cfd9e864802/nmstate-metrics/0.log" Mar 17 10:12:05 crc kubenswrapper[4813]: I0317 10:12:05.421753 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-k2x5n_0d0f7c9a-ffef-4d0b-ae31-0cfd9e864802/kube-rbac-proxy/0.log" Mar 17 10:12:05 crc kubenswrapper[4813]: I0317 10:12:05.512201 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-77p2p_acc2af0d-b40c-48d9-9b83-d0cf7755c37b/nmstate-operator/0.log" Mar 17 10:12:05 crc kubenswrapper[4813]: I0317 10:12:05.517414 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562372-jrqr7" event={"ID":"b7b6e6f5-528c-40fa-8f42-892a57e71fee","Type":"ContainerDied","Data":"eec39811a0873e08874c644be4af6ad6407357cf661b45c57b9124c4ad5da82b"} Mar 17 10:12:05 crc kubenswrapper[4813]: I0317 10:12:05.517449 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eec39811a0873e08874c644be4af6ad6407357cf661b45c57b9124c4ad5da82b" Mar 17 10:12:05 crc kubenswrapper[4813]: I0317 10:12:05.517506 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562372-jrqr7" Mar 17 10:12:05 crc kubenswrapper[4813]: I0317 10:12:05.572143 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562366-vpwp5"] Mar 17 10:12:05 crc kubenswrapper[4813]: I0317 10:12:05.581076 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562366-vpwp5"] Mar 17 10:12:05 crc kubenswrapper[4813]: I0317 10:12:05.600816 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-95n52_23256f06-964f-4f6f-bedf-6159a92a06d0/nmstate-webhook/0.log" Mar 17 10:12:06 crc kubenswrapper[4813]: I0317 10:12:06.740127 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e04f629c-67d3-4578-bb16-cf3656db7e77" path="/var/lib/kubelet/pods/e04f629c-67d3-4578-bb16-cf3656db7e77/volumes" Mar 17 10:12:13 crc kubenswrapper[4813]: I0317 10:12:13.731864 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:12:13 crc kubenswrapper[4813]: E0317 10:12:13.733219 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:12:28 crc kubenswrapper[4813]: I0317 10:12:28.746231 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:12:28 crc kubenswrapper[4813]: E0317 10:12:28.747244 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:12:33 crc kubenswrapper[4813]: I0317 10:12:33.097024 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-5cxrm_8503b49f-ae8d-4e0d-b3c6-30320e0e283a/kube-rbac-proxy/0.log" Mar 17 10:12:33 crc kubenswrapper[4813]: I0317 10:12:33.194144 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-5cxrm_8503b49f-ae8d-4e0d-b3c6-30320e0e283a/controller/0.log" Mar 17 10:12:33 crc kubenswrapper[4813]: I0317 10:12:33.324316 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-frr-files/0.log" Mar 17 10:12:33 crc kubenswrapper[4813]: I0317 10:12:33.493615 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-reloader/0.log" Mar 17 10:12:33 crc kubenswrapper[4813]: I0317 10:12:33.495005 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-frr-files/0.log" Mar 17 10:12:33 crc kubenswrapper[4813]: I0317 10:12:33.499868 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-reloader/0.log" Mar 17 10:12:33 crc kubenswrapper[4813]: I0317 10:12:33.510924 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-metrics/0.log" Mar 17 10:12:33 crc kubenswrapper[4813]: I0317 10:12:33.664206 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-reloader/0.log" Mar 17 10:12:33 crc kubenswrapper[4813]: I0317 10:12:33.680904 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-frr-files/0.log" Mar 17 10:12:33 crc kubenswrapper[4813]: I0317 10:12:33.726237 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-metrics/0.log" Mar 17 10:12:33 crc kubenswrapper[4813]: I0317 10:12:33.760579 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-metrics/0.log" Mar 17 10:12:33 crc kubenswrapper[4813]: I0317 10:12:33.896306 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-reloader/0.log" Mar 17 10:12:33 crc kubenswrapper[4813]: I0317 10:12:33.897700 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-frr-files/0.log" Mar 17 10:12:33 crc kubenswrapper[4813]: I0317 10:12:33.909237 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-metrics/0.log" Mar 17 10:12:33 crc kubenswrapper[4813]: I0317 10:12:33.937047 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/controller/0.log" Mar 17 10:12:34 crc kubenswrapper[4813]: I0317 10:12:34.080077 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/kube-rbac-proxy/0.log" Mar 17 10:12:34 crc kubenswrapper[4813]: I0317 10:12:34.082207 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/frr-metrics/0.log" Mar 17 10:12:34 crc kubenswrapper[4813]: I0317 10:12:34.124339 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/kube-rbac-proxy-frr/0.log" Mar 17 10:12:34 crc kubenswrapper[4813]: I0317 10:12:34.252216 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/reloader/0.log" Mar 17 10:12:34 crc kubenswrapper[4813]: I0317 10:12:34.372327 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-wjgfj_30c2f61b-c33a-49c4-bde0-ffe8593f29b8/frr-k8s-webhook-server/0.log" Mar 17 10:12:34 crc kubenswrapper[4813]: I0317 10:12:34.549112 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-56666fbc9b-dj55n_a63b7bd8-1a98-40d9-9f56-c7662d229b3d/manager/0.log" Mar 17 10:12:34 crc kubenswrapper[4813]: I0317 10:12:34.691269 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-55494b4d7-h6xch_f83986c8-b147-48b9-84ec-f09d01ddd8ce/webhook-server/0.log" Mar 17 10:12:34 crc kubenswrapper[4813]: I0317 10:12:34.824548 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hzk8t_ece7eee8-f37e-43f4-a762-986d7d1ae876/kube-rbac-proxy/0.log" Mar 17 10:12:35 crc kubenswrapper[4813]: I0317 10:12:35.344379 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hzk8t_ece7eee8-f37e-43f4-a762-986d7d1ae876/speaker/0.log" Mar 17 10:12:35 crc kubenswrapper[4813]: I0317 10:12:35.473187 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/frr/0.log" Mar 17 10:12:43 crc kubenswrapper[4813]: I0317 10:12:43.731069 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:12:43 crc kubenswrapper[4813]: E0317 10:12:43.733899 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:12:49 crc kubenswrapper[4813]: I0317 10:12:49.037827 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf_bb556790-2a96-412b-9847-3cd62873bfed/util/0.log" Mar 17 10:12:49 crc kubenswrapper[4813]: I0317 10:12:49.138655 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf_bb556790-2a96-412b-9847-3cd62873bfed/util/0.log" Mar 17 10:12:49 crc kubenswrapper[4813]: I0317 10:12:49.174905 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf_bb556790-2a96-412b-9847-3cd62873bfed/pull/0.log" Mar 17 10:12:49 crc kubenswrapper[4813]: I0317 10:12:49.231630 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf_bb556790-2a96-412b-9847-3cd62873bfed/pull/0.log" Mar 17 10:12:49 crc kubenswrapper[4813]: I0317 10:12:49.365078 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf_bb556790-2a96-412b-9847-3cd62873bfed/util/0.log" Mar 17 10:12:49 crc kubenswrapper[4813]: I0317 10:12:49.383026 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf_bb556790-2a96-412b-9847-3cd62873bfed/pull/0.log" Mar 17 10:12:49 crc kubenswrapper[4813]: I0317 10:12:49.394213 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf_bb556790-2a96-412b-9847-3cd62873bfed/extract/0.log" Mar 17 10:12:49 crc kubenswrapper[4813]: I0317 10:12:49.585422 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98_0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2/util/0.log" Mar 17 10:12:49 crc kubenswrapper[4813]: I0317 10:12:49.728169 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98_0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2/util/0.log" Mar 17 10:12:49 crc kubenswrapper[4813]: I0317 10:12:49.762375 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98_0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2/pull/0.log" Mar 17 10:12:49 crc kubenswrapper[4813]: I0317 10:12:49.766532 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98_0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2/pull/0.log" Mar 17 10:12:49 crc kubenswrapper[4813]: I0317 10:12:49.924889 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98_0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2/pull/0.log" Mar 17 10:12:49 crc kubenswrapper[4813]: I0317 10:12:49.933227 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98_0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2/extract/0.log" Mar 17 10:12:49 crc kubenswrapper[4813]: I0317 10:12:49.939178 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98_0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2/util/0.log" Mar 17 10:12:50 crc kubenswrapper[4813]: I0317 10:12:50.115137 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q2m6c_8840c430-4827-4da1-baec-38cff419ed36/extract-utilities/0.log" Mar 17 10:12:50 crc kubenswrapper[4813]: I0317 10:12:50.283993 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q2m6c_8840c430-4827-4da1-baec-38cff419ed36/extract-content/0.log" Mar 17 10:12:50 crc kubenswrapper[4813]: I0317 10:12:50.298216 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q2m6c_8840c430-4827-4da1-baec-38cff419ed36/extract-content/0.log" Mar 17 10:12:50 crc kubenswrapper[4813]: I0317 10:12:50.302881 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q2m6c_8840c430-4827-4da1-baec-38cff419ed36/extract-utilities/0.log" Mar 17 10:12:50 crc kubenswrapper[4813]: I0317 10:12:50.452272 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q2m6c_8840c430-4827-4da1-baec-38cff419ed36/extract-content/0.log" Mar 17 10:12:50 crc kubenswrapper[4813]: I0317 10:12:50.459913 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q2m6c_8840c430-4827-4da1-baec-38cff419ed36/extract-utilities/0.log" Mar 17 10:12:50 crc kubenswrapper[4813]: I0317 10:12:50.650547 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pr6tk_f8b71cf5-0aee-4d4e-a861-bb1bb52fe746/extract-utilities/0.log" Mar 17 10:12:50 crc kubenswrapper[4813]: I0317 10:12:50.676483 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q2m6c_8840c430-4827-4da1-baec-38cff419ed36/registry-server/0.log" Mar 17 10:12:50 crc kubenswrapper[4813]: I0317 10:12:50.881970 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pr6tk_f8b71cf5-0aee-4d4e-a861-bb1bb52fe746/extract-content/0.log" Mar 17 10:12:50 crc kubenswrapper[4813]: I0317 10:12:50.882143 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pr6tk_f8b71cf5-0aee-4d4e-a861-bb1bb52fe746/extract-utilities/0.log" Mar 17 10:12:50 crc kubenswrapper[4813]: I0317 10:12:50.892923 4813 scope.go:117] "RemoveContainer" containerID="69fa314a70f55e1127d1a0ae14d1cd2f3fa8f93264d3f915c843376809f754c7" Mar 17 10:12:50 crc kubenswrapper[4813]: I0317 10:12:50.924179 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pr6tk_f8b71cf5-0aee-4d4e-a861-bb1bb52fe746/extract-content/0.log" Mar 17 10:12:51 crc kubenswrapper[4813]: I0317 10:12:51.090642 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pr6tk_f8b71cf5-0aee-4d4e-a861-bb1bb52fe746/extract-content/0.log" Mar 17 10:12:51 crc kubenswrapper[4813]: I0317 10:12:51.096142 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pr6tk_f8b71cf5-0aee-4d4e-a861-bb1bb52fe746/extract-utilities/0.log" Mar 17 10:12:51 crc kubenswrapper[4813]: I0317 10:12:51.348022 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pr6tk_f8b71cf5-0aee-4d4e-a861-bb1bb52fe746/registry-server/0.log" Mar 17 10:12:51 crc kubenswrapper[4813]: I0317 10:12:51.376748 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-nw44l_6ca7f195-b5c8-4804-b95a-5ff53a637ac5/marketplace-operator/0.log" Mar 17 10:12:51 crc kubenswrapper[4813]: I0317 10:12:51.377556 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-54j2x_c7e12568-6cf2-4551-8aff-b330c8eda201/extract-utilities/0.log" Mar 17 10:12:51 crc kubenswrapper[4813]: I0317 10:12:51.519415 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-54j2x_c7e12568-6cf2-4551-8aff-b330c8eda201/extract-content/0.log" Mar 17 10:12:51 crc kubenswrapper[4813]: I0317 10:12:51.536348 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-54j2x_c7e12568-6cf2-4551-8aff-b330c8eda201/extract-content/0.log" Mar 17 10:12:51 crc kubenswrapper[4813]: I0317 10:12:51.536560 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-54j2x_c7e12568-6cf2-4551-8aff-b330c8eda201/extract-utilities/0.log" Mar 17 10:12:51 crc kubenswrapper[4813]: I0317 10:12:51.704531 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-54j2x_c7e12568-6cf2-4551-8aff-b330c8eda201/extract-content/0.log" Mar 17 10:12:51 crc kubenswrapper[4813]: I0317 10:12:51.717911 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-54j2x_c7e12568-6cf2-4551-8aff-b330c8eda201/extract-utilities/0.log" Mar 17 10:12:51 crc kubenswrapper[4813]: I0317 10:12:51.840741 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-54j2x_c7e12568-6cf2-4551-8aff-b330c8eda201/registry-server/0.log" Mar 17 10:12:51 crc kubenswrapper[4813]: I0317 10:12:51.881411 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vf8rq_4fc7a47b-182b-40dc-bf52-752b7434f448/extract-utilities/0.log" Mar 17 10:12:52 crc kubenswrapper[4813]: I0317 10:12:52.124893 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vf8rq_4fc7a47b-182b-40dc-bf52-752b7434f448/extract-utilities/0.log" Mar 17 10:12:52 crc kubenswrapper[4813]: I0317 10:12:52.128204 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vf8rq_4fc7a47b-182b-40dc-bf52-752b7434f448/extract-content/0.log" Mar 17 10:12:52 crc kubenswrapper[4813]: I0317 10:12:52.138006 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vf8rq_4fc7a47b-182b-40dc-bf52-752b7434f448/extract-content/0.log" Mar 17 10:12:52 crc kubenswrapper[4813]: I0317 10:12:52.316513 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vf8rq_4fc7a47b-182b-40dc-bf52-752b7434f448/extract-utilities/0.log" Mar 17 10:12:52 crc kubenswrapper[4813]: I0317 10:12:52.326980 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vf8rq_4fc7a47b-182b-40dc-bf52-752b7434f448/extract-content/0.log" Mar 17 10:12:52 crc kubenswrapper[4813]: I0317 10:12:52.779475 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vf8rq_4fc7a47b-182b-40dc-bf52-752b7434f448/registry-server/0.log" Mar 17 10:12:56 crc kubenswrapper[4813]: I0317 10:12:56.731370 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:12:56 crc kubenswrapper[4813]: E0317 10:12:56.732346 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:13:08 crc kubenswrapper[4813]: I0317 10:13:08.737007 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:13:08 crc kubenswrapper[4813]: E0317 10:13:08.737873 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:13:15 crc kubenswrapper[4813]: E0317 10:13:15.949201 4813 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.143:59404->38.102.83.143:40999: write tcp 38.102.83.143:59404->38.102.83.143:40999: write: broken pipe Mar 17 10:13:19 crc kubenswrapper[4813]: I0317 10:13:19.730457 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:13:19 crc kubenswrapper[4813]: E0317 10:13:19.731262 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:13:34 crc kubenswrapper[4813]: I0317 10:13:34.730586 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:13:34 crc kubenswrapper[4813]: E0317 10:13:34.731468 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:13:46 crc kubenswrapper[4813]: I0317 10:13:46.734999 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:13:46 crc kubenswrapper[4813]: E0317 10:13:46.735929 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:14:00 crc kubenswrapper[4813]: I0317 10:14:00.158811 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562374-sx7jn"] Mar 17 10:14:00 crc kubenswrapper[4813]: E0317 10:14:00.161230 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7b6e6f5-528c-40fa-8f42-892a57e71fee" containerName="oc" Mar 17 10:14:00 crc kubenswrapper[4813]: I0317 10:14:00.161252 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7b6e6f5-528c-40fa-8f42-892a57e71fee" containerName="oc" Mar 17 10:14:00 crc kubenswrapper[4813]: I0317 10:14:00.161482 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7b6e6f5-528c-40fa-8f42-892a57e71fee" containerName="oc" Mar 17 10:14:00 crc kubenswrapper[4813]: I0317 10:14:00.162258 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562374-sx7jn" Mar 17 10:14:00 crc kubenswrapper[4813]: I0317 10:14:00.171061 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 10:14:00 crc kubenswrapper[4813]: I0317 10:14:00.171320 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 10:14:00 crc kubenswrapper[4813]: I0317 10:14:00.171425 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 10:14:00 crc kubenswrapper[4813]: I0317 10:14:00.192839 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562374-sx7jn"] Mar 17 10:14:00 crc kubenswrapper[4813]: I0317 10:14:00.295924 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5hhg\" (UniqueName: \"kubernetes.io/projected/c4330a08-3146-4c6d-aa2e-ba94a5d7d24c-kube-api-access-l5hhg\") pod \"auto-csr-approver-29562374-sx7jn\" (UID: \"c4330a08-3146-4c6d-aa2e-ba94a5d7d24c\") " pod="openshift-infra/auto-csr-approver-29562374-sx7jn" Mar 17 10:14:00 crc kubenswrapper[4813]: I0317 10:14:00.397835 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5hhg\" (UniqueName: \"kubernetes.io/projected/c4330a08-3146-4c6d-aa2e-ba94a5d7d24c-kube-api-access-l5hhg\") pod \"auto-csr-approver-29562374-sx7jn\" (UID: \"c4330a08-3146-4c6d-aa2e-ba94a5d7d24c\") " pod="openshift-infra/auto-csr-approver-29562374-sx7jn" Mar 17 10:14:00 crc kubenswrapper[4813]: I0317 10:14:00.430229 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5hhg\" (UniqueName: \"kubernetes.io/projected/c4330a08-3146-4c6d-aa2e-ba94a5d7d24c-kube-api-access-l5hhg\") pod \"auto-csr-approver-29562374-sx7jn\" (UID: \"c4330a08-3146-4c6d-aa2e-ba94a5d7d24c\") " pod="openshift-infra/auto-csr-approver-29562374-sx7jn" Mar 17 10:14:00 crc kubenswrapper[4813]: I0317 10:14:00.501326 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562374-sx7jn" Mar 17 10:14:00 crc kubenswrapper[4813]: I0317 10:14:00.730759 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:14:00 crc kubenswrapper[4813]: E0317 10:14:00.731071 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:14:00 crc kubenswrapper[4813]: I0317 10:14:00.979705 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562374-sx7jn"] Mar 17 10:14:01 crc kubenswrapper[4813]: I0317 10:14:01.680500 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562374-sx7jn" event={"ID":"c4330a08-3146-4c6d-aa2e-ba94a5d7d24c","Type":"ContainerStarted","Data":"d954bcc0ec96ca44acebb9f131c536914566d66e53cb8d492a3014b11b9fe76a"} Mar 17 10:14:02 crc kubenswrapper[4813]: I0317 10:14:02.691975 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562374-sx7jn" event={"ID":"c4330a08-3146-4c6d-aa2e-ba94a5d7d24c","Type":"ContainerStarted","Data":"9b867a204e4e466a5ab442da560f36a8efd01a81a3e34a285b48127292737f1a"} Mar 17 10:14:02 crc kubenswrapper[4813]: I0317 10:14:02.713335 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562374-sx7jn" podStartSLOduration=1.542748987 podStartE2EDuration="2.713317008s" podCreationTimestamp="2026-03-17 10:14:00 +0000 UTC" firstStartedPulling="2026-03-17 10:14:01.010838899 +0000 UTC m=+3863.111642398" lastFinishedPulling="2026-03-17 10:14:02.18140692 +0000 UTC m=+3864.282210419" observedRunningTime="2026-03-17 10:14:02.704474502 +0000 UTC m=+3864.805278001" watchObservedRunningTime="2026-03-17 10:14:02.713317008 +0000 UTC m=+3864.814120507" Mar 17 10:14:03 crc kubenswrapper[4813]: I0317 10:14:03.704797 4813 generic.go:334] "Generic (PLEG): container finished" podID="c4330a08-3146-4c6d-aa2e-ba94a5d7d24c" containerID="9b867a204e4e466a5ab442da560f36a8efd01a81a3e34a285b48127292737f1a" exitCode=0 Mar 17 10:14:03 crc kubenswrapper[4813]: I0317 10:14:03.704866 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562374-sx7jn" event={"ID":"c4330a08-3146-4c6d-aa2e-ba94a5d7d24c","Type":"ContainerDied","Data":"9b867a204e4e466a5ab442da560f36a8efd01a81a3e34a285b48127292737f1a"} Mar 17 10:14:05 crc kubenswrapper[4813]: I0317 10:14:05.093446 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562374-sx7jn" Mar 17 10:14:05 crc kubenswrapper[4813]: I0317 10:14:05.189710 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5hhg\" (UniqueName: \"kubernetes.io/projected/c4330a08-3146-4c6d-aa2e-ba94a5d7d24c-kube-api-access-l5hhg\") pod \"c4330a08-3146-4c6d-aa2e-ba94a5d7d24c\" (UID: \"c4330a08-3146-4c6d-aa2e-ba94a5d7d24c\") " Mar 17 10:14:05 crc kubenswrapper[4813]: I0317 10:14:05.199018 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c4330a08-3146-4c6d-aa2e-ba94a5d7d24c-kube-api-access-l5hhg" (OuterVolumeSpecName: "kube-api-access-l5hhg") pod "c4330a08-3146-4c6d-aa2e-ba94a5d7d24c" (UID: "c4330a08-3146-4c6d-aa2e-ba94a5d7d24c"). InnerVolumeSpecName "kube-api-access-l5hhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:14:05 crc kubenswrapper[4813]: I0317 10:14:05.292034 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5hhg\" (UniqueName: \"kubernetes.io/projected/c4330a08-3146-4c6d-aa2e-ba94a5d7d24c-kube-api-access-l5hhg\") on node \"crc\" DevicePath \"\"" Mar 17 10:14:05 crc kubenswrapper[4813]: I0317 10:14:05.728610 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562374-sx7jn" event={"ID":"c4330a08-3146-4c6d-aa2e-ba94a5d7d24c","Type":"ContainerDied","Data":"d954bcc0ec96ca44acebb9f131c536914566d66e53cb8d492a3014b11b9fe76a"} Mar 17 10:14:05 crc kubenswrapper[4813]: I0317 10:14:05.728986 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d954bcc0ec96ca44acebb9f131c536914566d66e53cb8d492a3014b11b9fe76a" Mar 17 10:14:05 crc kubenswrapper[4813]: I0317 10:14:05.728706 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562374-sx7jn" Mar 17 10:14:05 crc kubenswrapper[4813]: I0317 10:14:05.838962 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562368-tkpms"] Mar 17 10:14:05 crc kubenswrapper[4813]: I0317 10:14:05.854124 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562368-tkpms"] Mar 17 10:14:06 crc kubenswrapper[4813]: I0317 10:14:06.743509 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39c26c40-ab53-43cf-93dd-14bd56d0dd6c" path="/var/lib/kubelet/pods/39c26c40-ab53-43cf-93dd-14bd56d0dd6c/volumes" Mar 17 10:14:11 crc kubenswrapper[4813]: I0317 10:14:11.730706 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:14:11 crc kubenswrapper[4813]: E0317 10:14:11.731794 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:14:25 crc kubenswrapper[4813]: I0317 10:14:25.731476 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:14:25 crc kubenswrapper[4813]: E0317 10:14:25.732282 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:14:34 crc kubenswrapper[4813]: E0317 10:14:34.241939 4813 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ed7b374_66e7_4391_9fa6_5d38d851d911.slice/crio-conmon-bd56a0dbbc97d78e263eec34ae6f377706150b613558a4d29ae98d4a3d9cbdda.scope\": RecentStats: unable to find data in memory cache]" Mar 17 10:14:35 crc kubenswrapper[4813]: I0317 10:14:35.030424 4813 generic.go:334] "Generic (PLEG): container finished" podID="9ed7b374-66e7-4391-9fa6-5d38d851d911" containerID="bd56a0dbbc97d78e263eec34ae6f377706150b613558a4d29ae98d4a3d9cbdda" exitCode=0 Mar 17 10:14:35 crc kubenswrapper[4813]: I0317 10:14:35.030581 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-d49m5/must-gather-676dh" event={"ID":"9ed7b374-66e7-4391-9fa6-5d38d851d911","Type":"ContainerDied","Data":"bd56a0dbbc97d78e263eec34ae6f377706150b613558a4d29ae98d4a3d9cbdda"} Mar 17 10:14:35 crc kubenswrapper[4813]: I0317 10:14:35.031448 4813 scope.go:117] "RemoveContainer" containerID="bd56a0dbbc97d78e263eec34ae6f377706150b613558a4d29ae98d4a3d9cbdda" Mar 17 10:14:35 crc kubenswrapper[4813]: I0317 10:14:35.142927 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-d49m5_must-gather-676dh_9ed7b374-66e7-4391-9fa6-5d38d851d911/gather/0.log" Mar 17 10:14:38 crc kubenswrapper[4813]: I0317 10:14:38.746391 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:14:38 crc kubenswrapper[4813]: E0317 10:14:38.747265 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:14:43 crc kubenswrapper[4813]: I0317 10:14:43.651251 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-d49m5/must-gather-676dh"] Mar 17 10:14:43 crc kubenswrapper[4813]: I0317 10:14:43.652142 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-d49m5/must-gather-676dh" podUID="9ed7b374-66e7-4391-9fa6-5d38d851d911" containerName="copy" containerID="cri-o://274968342ec150566781c5e622b6db7f033fc8d0ac42a9dbf038fb70d308b6ce" gracePeriod=2 Mar 17 10:14:43 crc kubenswrapper[4813]: I0317 10:14:43.663048 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-d49m5/must-gather-676dh"] Mar 17 10:14:44 crc kubenswrapper[4813]: I0317 10:14:44.127037 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-d49m5_must-gather-676dh_9ed7b374-66e7-4391-9fa6-5d38d851d911/copy/0.log" Mar 17 10:14:44 crc kubenswrapper[4813]: I0317 10:14:44.127864 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d49m5/must-gather-676dh" Mar 17 10:14:44 crc kubenswrapper[4813]: I0317 10:14:44.141807 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-d49m5_must-gather-676dh_9ed7b374-66e7-4391-9fa6-5d38d851d911/copy/0.log" Mar 17 10:14:44 crc kubenswrapper[4813]: I0317 10:14:44.142433 4813 generic.go:334] "Generic (PLEG): container finished" podID="9ed7b374-66e7-4391-9fa6-5d38d851d911" containerID="274968342ec150566781c5e622b6db7f033fc8d0ac42a9dbf038fb70d308b6ce" exitCode=143 Mar 17 10:14:44 crc kubenswrapper[4813]: I0317 10:14:44.142502 4813 scope.go:117] "RemoveContainer" containerID="274968342ec150566781c5e622b6db7f033fc8d0ac42a9dbf038fb70d308b6ce" Mar 17 10:14:44 crc kubenswrapper[4813]: I0317 10:14:44.142511 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-d49m5/must-gather-676dh" Mar 17 10:14:44 crc kubenswrapper[4813]: I0317 10:14:44.170005 4813 scope.go:117] "RemoveContainer" containerID="bd56a0dbbc97d78e263eec34ae6f377706150b613558a4d29ae98d4a3d9cbdda" Mar 17 10:14:44 crc kubenswrapper[4813]: I0317 10:14:44.247268 4813 scope.go:117] "RemoveContainer" containerID="274968342ec150566781c5e622b6db7f033fc8d0ac42a9dbf038fb70d308b6ce" Mar 17 10:14:44 crc kubenswrapper[4813]: E0317 10:14:44.247728 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"274968342ec150566781c5e622b6db7f033fc8d0ac42a9dbf038fb70d308b6ce\": container with ID starting with 274968342ec150566781c5e622b6db7f033fc8d0ac42a9dbf038fb70d308b6ce not found: ID does not exist" containerID="274968342ec150566781c5e622b6db7f033fc8d0ac42a9dbf038fb70d308b6ce" Mar 17 10:14:44 crc kubenswrapper[4813]: I0317 10:14:44.247768 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"274968342ec150566781c5e622b6db7f033fc8d0ac42a9dbf038fb70d308b6ce"} err="failed to get container status \"274968342ec150566781c5e622b6db7f033fc8d0ac42a9dbf038fb70d308b6ce\": rpc error: code = NotFound desc = could not find container \"274968342ec150566781c5e622b6db7f033fc8d0ac42a9dbf038fb70d308b6ce\": container with ID starting with 274968342ec150566781c5e622b6db7f033fc8d0ac42a9dbf038fb70d308b6ce not found: ID does not exist" Mar 17 10:14:44 crc kubenswrapper[4813]: I0317 10:14:44.247801 4813 scope.go:117] "RemoveContainer" containerID="bd56a0dbbc97d78e263eec34ae6f377706150b613558a4d29ae98d4a3d9cbdda" Mar 17 10:14:44 crc kubenswrapper[4813]: E0317 10:14:44.248206 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd56a0dbbc97d78e263eec34ae6f377706150b613558a4d29ae98d4a3d9cbdda\": container with ID starting with bd56a0dbbc97d78e263eec34ae6f377706150b613558a4d29ae98d4a3d9cbdda not found: ID does not exist" containerID="bd56a0dbbc97d78e263eec34ae6f377706150b613558a4d29ae98d4a3d9cbdda" Mar 17 10:14:44 crc kubenswrapper[4813]: I0317 10:14:44.248250 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd56a0dbbc97d78e263eec34ae6f377706150b613558a4d29ae98d4a3d9cbdda"} err="failed to get container status \"bd56a0dbbc97d78e263eec34ae6f377706150b613558a4d29ae98d4a3d9cbdda\": rpc error: code = NotFound desc = could not find container \"bd56a0dbbc97d78e263eec34ae6f377706150b613558a4d29ae98d4a3d9cbdda\": container with ID starting with bd56a0dbbc97d78e263eec34ae6f377706150b613558a4d29ae98d4a3d9cbdda not found: ID does not exist" Mar 17 10:14:44 crc kubenswrapper[4813]: I0317 10:14:44.295896 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9ed7b374-66e7-4391-9fa6-5d38d851d911-must-gather-output\") pod \"9ed7b374-66e7-4391-9fa6-5d38d851d911\" (UID: \"9ed7b374-66e7-4391-9fa6-5d38d851d911\") " Mar 17 10:14:44 crc kubenswrapper[4813]: I0317 10:14:44.296059 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zd6rs\" (UniqueName: \"kubernetes.io/projected/9ed7b374-66e7-4391-9fa6-5d38d851d911-kube-api-access-zd6rs\") pod \"9ed7b374-66e7-4391-9fa6-5d38d851d911\" (UID: \"9ed7b374-66e7-4391-9fa6-5d38d851d911\") " Mar 17 10:14:44 crc kubenswrapper[4813]: I0317 10:14:44.305422 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ed7b374-66e7-4391-9fa6-5d38d851d911-kube-api-access-zd6rs" (OuterVolumeSpecName: "kube-api-access-zd6rs") pod "9ed7b374-66e7-4391-9fa6-5d38d851d911" (UID: "9ed7b374-66e7-4391-9fa6-5d38d851d911"). InnerVolumeSpecName "kube-api-access-zd6rs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:14:44 crc kubenswrapper[4813]: I0317 10:14:44.397570 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zd6rs\" (UniqueName: \"kubernetes.io/projected/9ed7b374-66e7-4391-9fa6-5d38d851d911-kube-api-access-zd6rs\") on node \"crc\" DevicePath \"\"" Mar 17 10:14:44 crc kubenswrapper[4813]: I0317 10:14:44.466379 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9ed7b374-66e7-4391-9fa6-5d38d851d911-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "9ed7b374-66e7-4391-9fa6-5d38d851d911" (UID: "9ed7b374-66e7-4391-9fa6-5d38d851d911"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:14:44 crc kubenswrapper[4813]: I0317 10:14:44.499545 4813 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9ed7b374-66e7-4391-9fa6-5d38d851d911-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 17 10:14:44 crc kubenswrapper[4813]: I0317 10:14:44.743780 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ed7b374-66e7-4391-9fa6-5d38d851d911" path="/var/lib/kubelet/pods/9ed7b374-66e7-4391-9fa6-5d38d851d911/volumes" Mar 17 10:14:50 crc kubenswrapper[4813]: I0317 10:14:50.973268 4813 scope.go:117] "RemoveContainer" containerID="d5819ff4cd414054ad7e184c5b13093f822f73b7fea23519ebbf70e4503292a2" Mar 17 10:14:53 crc kubenswrapper[4813]: I0317 10:14:53.732719 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:14:53 crc kubenswrapper[4813]: E0317 10:14:53.733836 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:15:00 crc kubenswrapper[4813]: I0317 10:15:00.154820 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562375-bhw9t"] Mar 17 10:15:00 crc kubenswrapper[4813]: E0317 10:15:00.155684 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ed7b374-66e7-4391-9fa6-5d38d851d911" containerName="gather" Mar 17 10:15:00 crc kubenswrapper[4813]: I0317 10:15:00.155696 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ed7b374-66e7-4391-9fa6-5d38d851d911" containerName="gather" Mar 17 10:15:00 crc kubenswrapper[4813]: E0317 10:15:00.155712 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ed7b374-66e7-4391-9fa6-5d38d851d911" containerName="copy" Mar 17 10:15:00 crc kubenswrapper[4813]: I0317 10:15:00.155718 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ed7b374-66e7-4391-9fa6-5d38d851d911" containerName="copy" Mar 17 10:15:00 crc kubenswrapper[4813]: E0317 10:15:00.155732 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c4330a08-3146-4c6d-aa2e-ba94a5d7d24c" containerName="oc" Mar 17 10:15:00 crc kubenswrapper[4813]: I0317 10:15:00.155739 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="c4330a08-3146-4c6d-aa2e-ba94a5d7d24c" containerName="oc" Mar 17 10:15:00 crc kubenswrapper[4813]: I0317 10:15:00.155908 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="c4330a08-3146-4c6d-aa2e-ba94a5d7d24c" containerName="oc" Mar 17 10:15:00 crc kubenswrapper[4813]: I0317 10:15:00.155922 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ed7b374-66e7-4391-9fa6-5d38d851d911" containerName="gather" Mar 17 10:15:00 crc kubenswrapper[4813]: I0317 10:15:00.155935 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ed7b374-66e7-4391-9fa6-5d38d851d911" containerName="copy" Mar 17 10:15:00 crc kubenswrapper[4813]: I0317 10:15:00.156653 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562375-bhw9t" Mar 17 10:15:00 crc kubenswrapper[4813]: I0317 10:15:00.162140 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 17 10:15:00 crc kubenswrapper[4813]: I0317 10:15:00.162594 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 17 10:15:00 crc kubenswrapper[4813]: I0317 10:15:00.183510 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562375-bhw9t"] Mar 17 10:15:00 crc kubenswrapper[4813]: I0317 10:15:00.228821 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1f5aca0-361a-4f7d-a41f-5ac3f89ca099-config-volume\") pod \"collect-profiles-29562375-bhw9t\" (UID: \"e1f5aca0-361a-4f7d-a41f-5ac3f89ca099\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562375-bhw9t" Mar 17 10:15:00 crc kubenswrapper[4813]: I0317 10:15:00.228877 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1f5aca0-361a-4f7d-a41f-5ac3f89ca099-secret-volume\") pod \"collect-profiles-29562375-bhw9t\" (UID: \"e1f5aca0-361a-4f7d-a41f-5ac3f89ca099\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562375-bhw9t" Mar 17 10:15:00 crc kubenswrapper[4813]: I0317 10:15:00.228940 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hpcfh\" (UniqueName: \"kubernetes.io/projected/e1f5aca0-361a-4f7d-a41f-5ac3f89ca099-kube-api-access-hpcfh\") pod \"collect-profiles-29562375-bhw9t\" (UID: \"e1f5aca0-361a-4f7d-a41f-5ac3f89ca099\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562375-bhw9t" Mar 17 10:15:00 crc kubenswrapper[4813]: I0317 10:15:00.330785 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1f5aca0-361a-4f7d-a41f-5ac3f89ca099-secret-volume\") pod \"collect-profiles-29562375-bhw9t\" (UID: \"e1f5aca0-361a-4f7d-a41f-5ac3f89ca099\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562375-bhw9t" Mar 17 10:15:00 crc kubenswrapper[4813]: I0317 10:15:00.330892 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hpcfh\" (UniqueName: \"kubernetes.io/projected/e1f5aca0-361a-4f7d-a41f-5ac3f89ca099-kube-api-access-hpcfh\") pod \"collect-profiles-29562375-bhw9t\" (UID: \"e1f5aca0-361a-4f7d-a41f-5ac3f89ca099\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562375-bhw9t" Mar 17 10:15:00 crc kubenswrapper[4813]: I0317 10:15:00.331026 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1f5aca0-361a-4f7d-a41f-5ac3f89ca099-config-volume\") pod \"collect-profiles-29562375-bhw9t\" (UID: \"e1f5aca0-361a-4f7d-a41f-5ac3f89ca099\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562375-bhw9t" Mar 17 10:15:00 crc kubenswrapper[4813]: I0317 10:15:00.333396 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1f5aca0-361a-4f7d-a41f-5ac3f89ca099-config-volume\") pod \"collect-profiles-29562375-bhw9t\" (UID: \"e1f5aca0-361a-4f7d-a41f-5ac3f89ca099\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562375-bhw9t" Mar 17 10:15:00 crc kubenswrapper[4813]: I0317 10:15:00.339928 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1f5aca0-361a-4f7d-a41f-5ac3f89ca099-secret-volume\") pod \"collect-profiles-29562375-bhw9t\" (UID: \"e1f5aca0-361a-4f7d-a41f-5ac3f89ca099\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562375-bhw9t" Mar 17 10:15:00 crc kubenswrapper[4813]: I0317 10:15:00.354843 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hpcfh\" (UniqueName: \"kubernetes.io/projected/e1f5aca0-361a-4f7d-a41f-5ac3f89ca099-kube-api-access-hpcfh\") pod \"collect-profiles-29562375-bhw9t\" (UID: \"e1f5aca0-361a-4f7d-a41f-5ac3f89ca099\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29562375-bhw9t" Mar 17 10:15:00 crc kubenswrapper[4813]: I0317 10:15:00.494730 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562375-bhw9t" Mar 17 10:15:01 crc kubenswrapper[4813]: I0317 10:15:01.015463 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562375-bhw9t"] Mar 17 10:15:01 crc kubenswrapper[4813]: I0317 10:15:01.321997 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562375-bhw9t" event={"ID":"e1f5aca0-361a-4f7d-a41f-5ac3f89ca099","Type":"ContainerStarted","Data":"5a4d6ed91d78ff55c0ca72906456c59ceacb783b543f6b825e8bb79bfc778762"} Mar 17 10:15:01 crc kubenswrapper[4813]: I0317 10:15:01.322043 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562375-bhw9t" event={"ID":"e1f5aca0-361a-4f7d-a41f-5ac3f89ca099","Type":"ContainerStarted","Data":"71c7e69e48b61fafe04caf24e770b6cb21aecc86b38cbe085a8717edc0580365"} Mar 17 10:15:01 crc kubenswrapper[4813]: I0317 10:15:01.347091 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29562375-bhw9t" podStartSLOduration=1.347070259 podStartE2EDuration="1.347070259s" podCreationTimestamp="2026-03-17 10:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 10:15:01.334760795 +0000 UTC m=+3923.435564314" watchObservedRunningTime="2026-03-17 10:15:01.347070259 +0000 UTC m=+3923.447873788" Mar 17 10:15:02 crc kubenswrapper[4813]: I0317 10:15:02.330892 4813 generic.go:334] "Generic (PLEG): container finished" podID="e1f5aca0-361a-4f7d-a41f-5ac3f89ca099" containerID="5a4d6ed91d78ff55c0ca72906456c59ceacb783b543f6b825e8bb79bfc778762" exitCode=0 Mar 17 10:15:02 crc kubenswrapper[4813]: I0317 10:15:02.330936 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562375-bhw9t" event={"ID":"e1f5aca0-361a-4f7d-a41f-5ac3f89ca099","Type":"ContainerDied","Data":"5a4d6ed91d78ff55c0ca72906456c59ceacb783b543f6b825e8bb79bfc778762"} Mar 17 10:15:03 crc kubenswrapper[4813]: I0317 10:15:03.823924 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562375-bhw9t" Mar 17 10:15:03 crc kubenswrapper[4813]: I0317 10:15:03.896104 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hpcfh\" (UniqueName: \"kubernetes.io/projected/e1f5aca0-361a-4f7d-a41f-5ac3f89ca099-kube-api-access-hpcfh\") pod \"e1f5aca0-361a-4f7d-a41f-5ac3f89ca099\" (UID: \"e1f5aca0-361a-4f7d-a41f-5ac3f89ca099\") " Mar 17 10:15:03 crc kubenswrapper[4813]: I0317 10:15:03.896239 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1f5aca0-361a-4f7d-a41f-5ac3f89ca099-secret-volume\") pod \"e1f5aca0-361a-4f7d-a41f-5ac3f89ca099\" (UID: \"e1f5aca0-361a-4f7d-a41f-5ac3f89ca099\") " Mar 17 10:15:03 crc kubenswrapper[4813]: I0317 10:15:03.896400 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1f5aca0-361a-4f7d-a41f-5ac3f89ca099-config-volume\") pod \"e1f5aca0-361a-4f7d-a41f-5ac3f89ca099\" (UID: \"e1f5aca0-361a-4f7d-a41f-5ac3f89ca099\") " Mar 17 10:15:03 crc kubenswrapper[4813]: I0317 10:15:03.896961 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e1f5aca0-361a-4f7d-a41f-5ac3f89ca099-config-volume" (OuterVolumeSpecName: "config-volume") pod "e1f5aca0-361a-4f7d-a41f-5ac3f89ca099" (UID: "e1f5aca0-361a-4f7d-a41f-5ac3f89ca099"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 17 10:15:03 crc kubenswrapper[4813]: I0317 10:15:03.897376 4813 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e1f5aca0-361a-4f7d-a41f-5ac3f89ca099-config-volume\") on node \"crc\" DevicePath \"\"" Mar 17 10:15:03 crc kubenswrapper[4813]: I0317 10:15:03.901383 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1f5aca0-361a-4f7d-a41f-5ac3f89ca099-kube-api-access-hpcfh" (OuterVolumeSpecName: "kube-api-access-hpcfh") pod "e1f5aca0-361a-4f7d-a41f-5ac3f89ca099" (UID: "e1f5aca0-361a-4f7d-a41f-5ac3f89ca099"). InnerVolumeSpecName "kube-api-access-hpcfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:15:03 crc kubenswrapper[4813]: I0317 10:15:03.901829 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e1f5aca0-361a-4f7d-a41f-5ac3f89ca099-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e1f5aca0-361a-4f7d-a41f-5ac3f89ca099" (UID: "e1f5aca0-361a-4f7d-a41f-5ac3f89ca099"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 17 10:15:03 crc kubenswrapper[4813]: I0317 10:15:03.999818 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hpcfh\" (UniqueName: \"kubernetes.io/projected/e1f5aca0-361a-4f7d-a41f-5ac3f89ca099-kube-api-access-hpcfh\") on node \"crc\" DevicePath \"\"" Mar 17 10:15:03 crc kubenswrapper[4813]: I0317 10:15:03.999878 4813 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e1f5aca0-361a-4f7d-a41f-5ac3f89ca099-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 17 10:15:04 crc kubenswrapper[4813]: I0317 10:15:04.356585 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29562375-bhw9t" event={"ID":"e1f5aca0-361a-4f7d-a41f-5ac3f89ca099","Type":"ContainerDied","Data":"71c7e69e48b61fafe04caf24e770b6cb21aecc86b38cbe085a8717edc0580365"} Mar 17 10:15:04 crc kubenswrapper[4813]: I0317 10:15:04.356965 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="71c7e69e48b61fafe04caf24e770b6cb21aecc86b38cbe085a8717edc0580365" Mar 17 10:15:04 crc kubenswrapper[4813]: I0317 10:15:04.356706 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29562375-bhw9t" Mar 17 10:15:04 crc kubenswrapper[4813]: I0317 10:15:04.429926 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562330-qjkx6"] Mar 17 10:15:04 crc kubenswrapper[4813]: I0317 10:15:04.437054 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29562330-qjkx6"] Mar 17 10:15:04 crc kubenswrapper[4813]: I0317 10:15:04.741292 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47bda8a6-acb6-48df-b15d-1561c26df6cf" path="/var/lib/kubelet/pods/47bda8a6-acb6-48df-b15d-1561c26df6cf/volumes" Mar 17 10:15:06 crc kubenswrapper[4813]: I0317 10:15:06.731654 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:15:06 crc kubenswrapper[4813]: E0317 10:15:06.732475 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:15:17 crc kubenswrapper[4813]: I0317 10:15:17.731097 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:15:17 crc kubenswrapper[4813]: E0317 10:15:17.731925 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:15:28 crc kubenswrapper[4813]: I0317 10:15:28.975227 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2p2xn"] Mar 17 10:15:28 crc kubenswrapper[4813]: E0317 10:15:28.976883 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1f5aca0-361a-4f7d-a41f-5ac3f89ca099" containerName="collect-profiles" Mar 17 10:15:28 crc kubenswrapper[4813]: I0317 10:15:28.976916 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1f5aca0-361a-4f7d-a41f-5ac3f89ca099" containerName="collect-profiles" Mar 17 10:15:28 crc kubenswrapper[4813]: I0317 10:15:28.977381 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1f5aca0-361a-4f7d-a41f-5ac3f89ca099" containerName="collect-profiles" Mar 17 10:15:28 crc kubenswrapper[4813]: I0317 10:15:28.980828 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2p2xn" Mar 17 10:15:28 crc kubenswrapper[4813]: I0317 10:15:28.990536 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2p2xn"] Mar 17 10:15:29 crc kubenswrapper[4813]: I0317 10:15:29.143093 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-88ll6\" (UniqueName: \"kubernetes.io/projected/cc2b75cc-64d3-411c-a6c7-96ee15e03834-kube-api-access-88ll6\") pod \"redhat-operators-2p2xn\" (UID: \"cc2b75cc-64d3-411c-a6c7-96ee15e03834\") " pod="openshift-marketplace/redhat-operators-2p2xn" Mar 17 10:15:29 crc kubenswrapper[4813]: I0317 10:15:29.143211 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc2b75cc-64d3-411c-a6c7-96ee15e03834-catalog-content\") pod \"redhat-operators-2p2xn\" (UID: \"cc2b75cc-64d3-411c-a6c7-96ee15e03834\") " pod="openshift-marketplace/redhat-operators-2p2xn" Mar 17 10:15:29 crc kubenswrapper[4813]: I0317 10:15:29.143317 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc2b75cc-64d3-411c-a6c7-96ee15e03834-utilities\") pod \"redhat-operators-2p2xn\" (UID: \"cc2b75cc-64d3-411c-a6c7-96ee15e03834\") " pod="openshift-marketplace/redhat-operators-2p2xn" Mar 17 10:15:29 crc kubenswrapper[4813]: I0317 10:15:29.245042 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc2b75cc-64d3-411c-a6c7-96ee15e03834-catalog-content\") pod \"redhat-operators-2p2xn\" (UID: \"cc2b75cc-64d3-411c-a6c7-96ee15e03834\") " pod="openshift-marketplace/redhat-operators-2p2xn" Mar 17 10:15:29 crc kubenswrapper[4813]: I0317 10:15:29.245144 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc2b75cc-64d3-411c-a6c7-96ee15e03834-utilities\") pod \"redhat-operators-2p2xn\" (UID: \"cc2b75cc-64d3-411c-a6c7-96ee15e03834\") " pod="openshift-marketplace/redhat-operators-2p2xn" Mar 17 10:15:29 crc kubenswrapper[4813]: I0317 10:15:29.245193 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-88ll6\" (UniqueName: \"kubernetes.io/projected/cc2b75cc-64d3-411c-a6c7-96ee15e03834-kube-api-access-88ll6\") pod \"redhat-operators-2p2xn\" (UID: \"cc2b75cc-64d3-411c-a6c7-96ee15e03834\") " pod="openshift-marketplace/redhat-operators-2p2xn" Mar 17 10:15:29 crc kubenswrapper[4813]: I0317 10:15:29.245874 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc2b75cc-64d3-411c-a6c7-96ee15e03834-catalog-content\") pod \"redhat-operators-2p2xn\" (UID: \"cc2b75cc-64d3-411c-a6c7-96ee15e03834\") " pod="openshift-marketplace/redhat-operators-2p2xn" Mar 17 10:15:29 crc kubenswrapper[4813]: I0317 10:15:29.246082 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc2b75cc-64d3-411c-a6c7-96ee15e03834-utilities\") pod \"redhat-operators-2p2xn\" (UID: \"cc2b75cc-64d3-411c-a6c7-96ee15e03834\") " pod="openshift-marketplace/redhat-operators-2p2xn" Mar 17 10:15:29 crc kubenswrapper[4813]: I0317 10:15:29.265514 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-88ll6\" (UniqueName: \"kubernetes.io/projected/cc2b75cc-64d3-411c-a6c7-96ee15e03834-kube-api-access-88ll6\") pod \"redhat-operators-2p2xn\" (UID: \"cc2b75cc-64d3-411c-a6c7-96ee15e03834\") " pod="openshift-marketplace/redhat-operators-2p2xn" Mar 17 10:15:29 crc kubenswrapper[4813]: I0317 10:15:29.319168 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2p2xn" Mar 17 10:15:29 crc kubenswrapper[4813]: I0317 10:15:29.764389 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2p2xn"] Mar 17 10:15:30 crc kubenswrapper[4813]: I0317 10:15:30.668801 4813 generic.go:334] "Generic (PLEG): container finished" podID="cc2b75cc-64d3-411c-a6c7-96ee15e03834" containerID="d8b01ff6fed683179bf73472bdb780f7c3fedfafcd33493d36ba58ec8c52c6a2" exitCode=0 Mar 17 10:15:30 crc kubenswrapper[4813]: I0317 10:15:30.668911 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2p2xn" event={"ID":"cc2b75cc-64d3-411c-a6c7-96ee15e03834","Type":"ContainerDied","Data":"d8b01ff6fed683179bf73472bdb780f7c3fedfafcd33493d36ba58ec8c52c6a2"} Mar 17 10:15:30 crc kubenswrapper[4813]: I0317 10:15:30.669169 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2p2xn" event={"ID":"cc2b75cc-64d3-411c-a6c7-96ee15e03834","Type":"ContainerStarted","Data":"85a0f8ec76a63434ad139723cc3324a059047476a2869079c10c26409ccbeb8e"} Mar 17 10:15:30 crc kubenswrapper[4813]: I0317 10:15:30.671769 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 10:15:32 crc kubenswrapper[4813]: I0317 10:15:32.694563 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2p2xn" event={"ID":"cc2b75cc-64d3-411c-a6c7-96ee15e03834","Type":"ContainerStarted","Data":"9f0dca4ac39d04ea8f26144824be8bbad98b18f2bfb41bc2378ba8f3dcfea2b7"} Mar 17 10:15:32 crc kubenswrapper[4813]: I0317 10:15:32.735036 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:15:32 crc kubenswrapper[4813]: E0317 10:15:32.735563 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:15:35 crc kubenswrapper[4813]: I0317 10:15:35.734113 4813 generic.go:334] "Generic (PLEG): container finished" podID="cc2b75cc-64d3-411c-a6c7-96ee15e03834" containerID="9f0dca4ac39d04ea8f26144824be8bbad98b18f2bfb41bc2378ba8f3dcfea2b7" exitCode=0 Mar 17 10:15:35 crc kubenswrapper[4813]: I0317 10:15:35.734791 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2p2xn" event={"ID":"cc2b75cc-64d3-411c-a6c7-96ee15e03834","Type":"ContainerDied","Data":"9f0dca4ac39d04ea8f26144824be8bbad98b18f2bfb41bc2378ba8f3dcfea2b7"} Mar 17 10:15:36 crc kubenswrapper[4813]: I0317 10:15:36.748940 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2p2xn" event={"ID":"cc2b75cc-64d3-411c-a6c7-96ee15e03834","Type":"ContainerStarted","Data":"7b82df1bb5e46ca33ff1d88a59fdac423975b82a728e0be20b2a3c34732b9be8"} Mar 17 10:15:36 crc kubenswrapper[4813]: I0317 10:15:36.782378 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2p2xn" podStartSLOduration=3.32177397 podStartE2EDuration="8.782346782s" podCreationTimestamp="2026-03-17 10:15:28 +0000 UTC" firstStartedPulling="2026-03-17 10:15:30.671209375 +0000 UTC m=+3952.772012914" lastFinishedPulling="2026-03-17 10:15:36.131782217 +0000 UTC m=+3958.232585726" observedRunningTime="2026-03-17 10:15:36.774094685 +0000 UTC m=+3958.874898204" watchObservedRunningTime="2026-03-17 10:15:36.782346782 +0000 UTC m=+3958.883150321" Mar 17 10:15:39 crc kubenswrapper[4813]: I0317 10:15:39.319432 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2p2xn" Mar 17 10:15:39 crc kubenswrapper[4813]: I0317 10:15:39.319881 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2p2xn" Mar 17 10:15:40 crc kubenswrapper[4813]: I0317 10:15:40.380716 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2p2xn" podUID="cc2b75cc-64d3-411c-a6c7-96ee15e03834" containerName="registry-server" probeResult="failure" output=< Mar 17 10:15:40 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Mar 17 10:15:40 crc kubenswrapper[4813]: > Mar 17 10:15:44 crc kubenswrapper[4813]: I0317 10:15:44.731294 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:15:44 crc kubenswrapper[4813]: E0317 10:15:44.734244 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:15:49 crc kubenswrapper[4813]: I0317 10:15:49.406346 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2p2xn" Mar 17 10:15:49 crc kubenswrapper[4813]: I0317 10:15:49.500074 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2p2xn" Mar 17 10:15:49 crc kubenswrapper[4813]: I0317 10:15:49.676806 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2p2xn"] Mar 17 10:15:50 crc kubenswrapper[4813]: I0317 10:15:50.902900 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2p2xn" podUID="cc2b75cc-64d3-411c-a6c7-96ee15e03834" containerName="registry-server" containerID="cri-o://7b82df1bb5e46ca33ff1d88a59fdac423975b82a728e0be20b2a3c34732b9be8" gracePeriod=2 Mar 17 10:15:51 crc kubenswrapper[4813]: I0317 10:15:51.108912 4813 scope.go:117] "RemoveContainer" containerID="07714735dd2134ad863dcb2368769844bf3220bfb769ca505a47f92572e07e2f" Mar 17 10:15:51 crc kubenswrapper[4813]: I0317 10:15:51.134523 4813 scope.go:117] "RemoveContainer" containerID="f2d9102e40557554cd79652b1df0de5876a52bf58b1793baf2ffc6a404a7548b" Mar 17 10:15:51 crc kubenswrapper[4813]: I0317 10:15:51.325559 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2p2xn" Mar 17 10:15:51 crc kubenswrapper[4813]: I0317 10:15:51.364120 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-88ll6\" (UniqueName: \"kubernetes.io/projected/cc2b75cc-64d3-411c-a6c7-96ee15e03834-kube-api-access-88ll6\") pod \"cc2b75cc-64d3-411c-a6c7-96ee15e03834\" (UID: \"cc2b75cc-64d3-411c-a6c7-96ee15e03834\") " Mar 17 10:15:51 crc kubenswrapper[4813]: I0317 10:15:51.364293 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc2b75cc-64d3-411c-a6c7-96ee15e03834-utilities\") pod \"cc2b75cc-64d3-411c-a6c7-96ee15e03834\" (UID: \"cc2b75cc-64d3-411c-a6c7-96ee15e03834\") " Mar 17 10:15:51 crc kubenswrapper[4813]: I0317 10:15:51.364377 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc2b75cc-64d3-411c-a6c7-96ee15e03834-catalog-content\") pod \"cc2b75cc-64d3-411c-a6c7-96ee15e03834\" (UID: \"cc2b75cc-64d3-411c-a6c7-96ee15e03834\") " Mar 17 10:15:51 crc kubenswrapper[4813]: I0317 10:15:51.365843 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc2b75cc-64d3-411c-a6c7-96ee15e03834-utilities" (OuterVolumeSpecName: "utilities") pod "cc2b75cc-64d3-411c-a6c7-96ee15e03834" (UID: "cc2b75cc-64d3-411c-a6c7-96ee15e03834"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:15:51 crc kubenswrapper[4813]: I0317 10:15:51.373248 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc2b75cc-64d3-411c-a6c7-96ee15e03834-kube-api-access-88ll6" (OuterVolumeSpecName: "kube-api-access-88ll6") pod "cc2b75cc-64d3-411c-a6c7-96ee15e03834" (UID: "cc2b75cc-64d3-411c-a6c7-96ee15e03834"). InnerVolumeSpecName "kube-api-access-88ll6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:15:51 crc kubenswrapper[4813]: I0317 10:15:51.467184 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-88ll6\" (UniqueName: \"kubernetes.io/projected/cc2b75cc-64d3-411c-a6c7-96ee15e03834-kube-api-access-88ll6\") on node \"crc\" DevicePath \"\"" Mar 17 10:15:51 crc kubenswrapper[4813]: I0317 10:15:51.467216 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc2b75cc-64d3-411c-a6c7-96ee15e03834-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 10:15:51 crc kubenswrapper[4813]: I0317 10:15:51.494838 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc2b75cc-64d3-411c-a6c7-96ee15e03834-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc2b75cc-64d3-411c-a6c7-96ee15e03834" (UID: "cc2b75cc-64d3-411c-a6c7-96ee15e03834"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:15:51 crc kubenswrapper[4813]: I0317 10:15:51.569142 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc2b75cc-64d3-411c-a6c7-96ee15e03834-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 10:15:51 crc kubenswrapper[4813]: I0317 10:15:51.918958 4813 generic.go:334] "Generic (PLEG): container finished" podID="cc2b75cc-64d3-411c-a6c7-96ee15e03834" containerID="7b82df1bb5e46ca33ff1d88a59fdac423975b82a728e0be20b2a3c34732b9be8" exitCode=0 Mar 17 10:15:51 crc kubenswrapper[4813]: I0317 10:15:51.919103 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2p2xn" event={"ID":"cc2b75cc-64d3-411c-a6c7-96ee15e03834","Type":"ContainerDied","Data":"7b82df1bb5e46ca33ff1d88a59fdac423975b82a728e0be20b2a3c34732b9be8"} Mar 17 10:15:51 crc kubenswrapper[4813]: I0317 10:15:51.919501 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2p2xn" event={"ID":"cc2b75cc-64d3-411c-a6c7-96ee15e03834","Type":"ContainerDied","Data":"85a0f8ec76a63434ad139723cc3324a059047476a2869079c10c26409ccbeb8e"} Mar 17 10:15:51 crc kubenswrapper[4813]: I0317 10:15:51.919539 4813 scope.go:117] "RemoveContainer" containerID="7b82df1bb5e46ca33ff1d88a59fdac423975b82a728e0be20b2a3c34732b9be8" Mar 17 10:15:51 crc kubenswrapper[4813]: I0317 10:15:51.919164 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2p2xn" Mar 17 10:15:51 crc kubenswrapper[4813]: I0317 10:15:51.956382 4813 scope.go:117] "RemoveContainer" containerID="9f0dca4ac39d04ea8f26144824be8bbad98b18f2bfb41bc2378ba8f3dcfea2b7" Mar 17 10:15:51 crc kubenswrapper[4813]: I0317 10:15:51.986888 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2p2xn"] Mar 17 10:15:51 crc kubenswrapper[4813]: I0317 10:15:51.995335 4813 scope.go:117] "RemoveContainer" containerID="d8b01ff6fed683179bf73472bdb780f7c3fedfafcd33493d36ba58ec8c52c6a2" Mar 17 10:15:52 crc kubenswrapper[4813]: I0317 10:15:52.003057 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2p2xn"] Mar 17 10:15:52 crc kubenswrapper[4813]: I0317 10:15:52.026447 4813 scope.go:117] "RemoveContainer" containerID="7b82df1bb5e46ca33ff1d88a59fdac423975b82a728e0be20b2a3c34732b9be8" Mar 17 10:15:52 crc kubenswrapper[4813]: E0317 10:15:52.027080 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b82df1bb5e46ca33ff1d88a59fdac423975b82a728e0be20b2a3c34732b9be8\": container with ID starting with 7b82df1bb5e46ca33ff1d88a59fdac423975b82a728e0be20b2a3c34732b9be8 not found: ID does not exist" containerID="7b82df1bb5e46ca33ff1d88a59fdac423975b82a728e0be20b2a3c34732b9be8" Mar 17 10:15:52 crc kubenswrapper[4813]: I0317 10:15:52.027142 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b82df1bb5e46ca33ff1d88a59fdac423975b82a728e0be20b2a3c34732b9be8"} err="failed to get container status \"7b82df1bb5e46ca33ff1d88a59fdac423975b82a728e0be20b2a3c34732b9be8\": rpc error: code = NotFound desc = could not find container \"7b82df1bb5e46ca33ff1d88a59fdac423975b82a728e0be20b2a3c34732b9be8\": container with ID starting with 7b82df1bb5e46ca33ff1d88a59fdac423975b82a728e0be20b2a3c34732b9be8 not found: ID does not exist" Mar 17 10:15:52 crc kubenswrapper[4813]: I0317 10:15:52.027181 4813 scope.go:117] "RemoveContainer" containerID="9f0dca4ac39d04ea8f26144824be8bbad98b18f2bfb41bc2378ba8f3dcfea2b7" Mar 17 10:15:52 crc kubenswrapper[4813]: E0317 10:15:52.027676 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f0dca4ac39d04ea8f26144824be8bbad98b18f2bfb41bc2378ba8f3dcfea2b7\": container with ID starting with 9f0dca4ac39d04ea8f26144824be8bbad98b18f2bfb41bc2378ba8f3dcfea2b7 not found: ID does not exist" containerID="9f0dca4ac39d04ea8f26144824be8bbad98b18f2bfb41bc2378ba8f3dcfea2b7" Mar 17 10:15:52 crc kubenswrapper[4813]: I0317 10:15:52.027708 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f0dca4ac39d04ea8f26144824be8bbad98b18f2bfb41bc2378ba8f3dcfea2b7"} err="failed to get container status \"9f0dca4ac39d04ea8f26144824be8bbad98b18f2bfb41bc2378ba8f3dcfea2b7\": rpc error: code = NotFound desc = could not find container \"9f0dca4ac39d04ea8f26144824be8bbad98b18f2bfb41bc2378ba8f3dcfea2b7\": container with ID starting with 9f0dca4ac39d04ea8f26144824be8bbad98b18f2bfb41bc2378ba8f3dcfea2b7 not found: ID does not exist" Mar 17 10:15:52 crc kubenswrapper[4813]: I0317 10:15:52.027734 4813 scope.go:117] "RemoveContainer" containerID="d8b01ff6fed683179bf73472bdb780f7c3fedfafcd33493d36ba58ec8c52c6a2" Mar 17 10:15:52 crc kubenswrapper[4813]: E0317 10:15:52.028206 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8b01ff6fed683179bf73472bdb780f7c3fedfafcd33493d36ba58ec8c52c6a2\": container with ID starting with d8b01ff6fed683179bf73472bdb780f7c3fedfafcd33493d36ba58ec8c52c6a2 not found: ID does not exist" containerID="d8b01ff6fed683179bf73472bdb780f7c3fedfafcd33493d36ba58ec8c52c6a2" Mar 17 10:15:52 crc kubenswrapper[4813]: I0317 10:15:52.028252 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8b01ff6fed683179bf73472bdb780f7c3fedfafcd33493d36ba58ec8c52c6a2"} err="failed to get container status \"d8b01ff6fed683179bf73472bdb780f7c3fedfafcd33493d36ba58ec8c52c6a2\": rpc error: code = NotFound desc = could not find container \"d8b01ff6fed683179bf73472bdb780f7c3fedfafcd33493d36ba58ec8c52c6a2\": container with ID starting with d8b01ff6fed683179bf73472bdb780f7c3fedfafcd33493d36ba58ec8c52c6a2 not found: ID does not exist" Mar 17 10:15:52 crc kubenswrapper[4813]: I0317 10:15:52.764877 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc2b75cc-64d3-411c-a6c7-96ee15e03834" path="/var/lib/kubelet/pods/cc2b75cc-64d3-411c-a6c7-96ee15e03834/volumes" Mar 17 10:15:57 crc kubenswrapper[4813]: I0317 10:15:57.731993 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:15:57 crc kubenswrapper[4813]: E0317 10:15:57.733209 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:16:00 crc kubenswrapper[4813]: I0317 10:16:00.213167 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562376-c2ztw"] Mar 17 10:16:00 crc kubenswrapper[4813]: E0317 10:16:00.213812 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc2b75cc-64d3-411c-a6c7-96ee15e03834" containerName="extract-content" Mar 17 10:16:00 crc kubenswrapper[4813]: I0317 10:16:00.213825 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc2b75cc-64d3-411c-a6c7-96ee15e03834" containerName="extract-content" Mar 17 10:16:00 crc kubenswrapper[4813]: E0317 10:16:00.213852 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc2b75cc-64d3-411c-a6c7-96ee15e03834" containerName="extract-utilities" Mar 17 10:16:00 crc kubenswrapper[4813]: I0317 10:16:00.213860 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc2b75cc-64d3-411c-a6c7-96ee15e03834" containerName="extract-utilities" Mar 17 10:16:00 crc kubenswrapper[4813]: E0317 10:16:00.213871 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc2b75cc-64d3-411c-a6c7-96ee15e03834" containerName="registry-server" Mar 17 10:16:00 crc kubenswrapper[4813]: I0317 10:16:00.213877 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc2b75cc-64d3-411c-a6c7-96ee15e03834" containerName="registry-server" Mar 17 10:16:00 crc kubenswrapper[4813]: I0317 10:16:00.214079 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc2b75cc-64d3-411c-a6c7-96ee15e03834" containerName="registry-server" Mar 17 10:16:00 crc kubenswrapper[4813]: I0317 10:16:00.214712 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562376-c2ztw" Mar 17 10:16:00 crc kubenswrapper[4813]: I0317 10:16:00.220107 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 10:16:00 crc kubenswrapper[4813]: I0317 10:16:00.220316 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 10:16:00 crc kubenswrapper[4813]: I0317 10:16:00.220422 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 10:16:00 crc kubenswrapper[4813]: I0317 10:16:00.223803 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562376-c2ztw"] Mar 17 10:16:00 crc kubenswrapper[4813]: I0317 10:16:00.371863 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nck82\" (UniqueName: \"kubernetes.io/projected/8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049-kube-api-access-nck82\") pod \"auto-csr-approver-29562376-c2ztw\" (UID: \"8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049\") " pod="openshift-infra/auto-csr-approver-29562376-c2ztw" Mar 17 10:16:00 crc kubenswrapper[4813]: I0317 10:16:00.473389 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nck82\" (UniqueName: \"kubernetes.io/projected/8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049-kube-api-access-nck82\") pod \"auto-csr-approver-29562376-c2ztw\" (UID: \"8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049\") " pod="openshift-infra/auto-csr-approver-29562376-c2ztw" Mar 17 10:16:00 crc kubenswrapper[4813]: I0317 10:16:00.499645 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nck82\" (UniqueName: \"kubernetes.io/projected/8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049-kube-api-access-nck82\") pod \"auto-csr-approver-29562376-c2ztw\" (UID: \"8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049\") " pod="openshift-infra/auto-csr-approver-29562376-c2ztw" Mar 17 10:16:00 crc kubenswrapper[4813]: I0317 10:16:00.550815 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562376-c2ztw" Mar 17 10:16:01 crc kubenswrapper[4813]: I0317 10:16:01.146228 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562376-c2ztw"] Mar 17 10:16:01 crc kubenswrapper[4813]: I0317 10:16:01.427290 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-fmd7w" Mar 17 10:16:02 crc kubenswrapper[4813]: I0317 10:16:02.042298 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562376-c2ztw" event={"ID":"8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049","Type":"ContainerStarted","Data":"57cc43ad88e5b85ef6c4a703ec5ff412392fdc1e315e915211355751d88d22d4"} Mar 17 10:16:03 crc kubenswrapper[4813]: I0317 10:16:03.060255 4813 generic.go:334] "Generic (PLEG): container finished" podID="8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049" containerID="e97424096bd14a70cdf2e53cd4c147471fe3dfd25e61c7104ecb286b2c55d8cd" exitCode=0 Mar 17 10:16:03 crc kubenswrapper[4813]: I0317 10:16:03.060316 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562376-c2ztw" event={"ID":"8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049","Type":"ContainerDied","Data":"e97424096bd14a70cdf2e53cd4c147471fe3dfd25e61c7104ecb286b2c55d8cd"} Mar 17 10:16:04 crc kubenswrapper[4813]: I0317 10:16:04.520183 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562376-c2ztw" Mar 17 10:16:04 crc kubenswrapper[4813]: I0317 10:16:04.550937 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nck82\" (UniqueName: \"kubernetes.io/projected/8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049-kube-api-access-nck82\") pod \"8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049\" (UID: \"8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049\") " Mar 17 10:16:04 crc kubenswrapper[4813]: I0317 10:16:04.559462 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049-kube-api-access-nck82" (OuterVolumeSpecName: "kube-api-access-nck82") pod "8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049" (UID: "8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049"). InnerVolumeSpecName "kube-api-access-nck82". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:16:04 crc kubenswrapper[4813]: I0317 10:16:04.653712 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nck82\" (UniqueName: \"kubernetes.io/projected/8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049-kube-api-access-nck82\") on node \"crc\" DevicePath \"\"" Mar 17 10:16:05 crc kubenswrapper[4813]: I0317 10:16:05.090762 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562376-c2ztw" event={"ID":"8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049","Type":"ContainerDied","Data":"57cc43ad88e5b85ef6c4a703ec5ff412392fdc1e315e915211355751d88d22d4"} Mar 17 10:16:05 crc kubenswrapper[4813]: I0317 10:16:05.090803 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57cc43ad88e5b85ef6c4a703ec5ff412392fdc1e315e915211355751d88d22d4" Mar 17 10:16:05 crc kubenswrapper[4813]: I0317 10:16:05.090861 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562376-c2ztw" Mar 17 10:16:05 crc kubenswrapper[4813]: I0317 10:16:05.611417 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562370-8rqq9"] Mar 17 10:16:05 crc kubenswrapper[4813]: I0317 10:16:05.623874 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562370-8rqq9"] Mar 17 10:16:06 crc kubenswrapper[4813]: I0317 10:16:06.746125 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a52ad15e-4714-47c1-87e7-b254fdd2d975" path="/var/lib/kubelet/pods/a52ad15e-4714-47c1-87e7-b254fdd2d975/volumes" Mar 17 10:16:11 crc kubenswrapper[4813]: I0317 10:16:11.738989 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:16:11 crc kubenswrapper[4813]: E0317 10:16:11.744834 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:16:22 crc kubenswrapper[4813]: I0317 10:16:22.731208 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:16:23 crc kubenswrapper[4813]: I0317 10:16:23.316558 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerStarted","Data":"38fdc8059c00804eb8afc6b3d9dac72a950b31fe31c73eb825628373cb5b059d"} Mar 17 10:16:51 crc kubenswrapper[4813]: I0317 10:16:51.240157 4813 scope.go:117] "RemoveContainer" containerID="2029cda7ecc42328ab03599a86967a8c885a5bfeddfb90555b49907725994eb8" Mar 17 10:16:51 crc kubenswrapper[4813]: I0317 10:16:51.274866 4813 scope.go:117] "RemoveContainer" containerID="03a45c4015d4a93a56227694dc938a468d8eebef6e657e2ff4bd8460561282c1" Mar 17 10:17:26 crc kubenswrapper[4813]: I0317 10:17:26.369236 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mvc44"] Mar 17 10:17:26 crc kubenswrapper[4813]: E0317 10:17:26.370163 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049" containerName="oc" Mar 17 10:17:26 crc kubenswrapper[4813]: I0317 10:17:26.370179 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049" containerName="oc" Mar 17 10:17:26 crc kubenswrapper[4813]: I0317 10:17:26.370421 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049" containerName="oc" Mar 17 10:17:26 crc kubenswrapper[4813]: I0317 10:17:26.372016 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mvc44" Mar 17 10:17:26 crc kubenswrapper[4813]: I0317 10:17:26.432103 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mvc44"] Mar 17 10:17:26 crc kubenswrapper[4813]: I0317 10:17:26.542091 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6-catalog-content\") pod \"community-operators-mvc44\" (UID: \"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6\") " pod="openshift-marketplace/community-operators-mvc44" Mar 17 10:17:26 crc kubenswrapper[4813]: I0317 10:17:26.542488 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ql7zj\" (UniqueName: \"kubernetes.io/projected/b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6-kube-api-access-ql7zj\") pod \"community-operators-mvc44\" (UID: \"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6\") " pod="openshift-marketplace/community-operators-mvc44" Mar 17 10:17:26 crc kubenswrapper[4813]: I0317 10:17:26.542644 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6-utilities\") pod \"community-operators-mvc44\" (UID: \"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6\") " pod="openshift-marketplace/community-operators-mvc44" Mar 17 10:17:26 crc kubenswrapper[4813]: I0317 10:17:26.644652 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6-catalog-content\") pod \"community-operators-mvc44\" (UID: \"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6\") " pod="openshift-marketplace/community-operators-mvc44" Mar 17 10:17:26 crc kubenswrapper[4813]: I0317 10:17:26.644782 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ql7zj\" (UniqueName: \"kubernetes.io/projected/b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6-kube-api-access-ql7zj\") pod \"community-operators-mvc44\" (UID: \"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6\") " pod="openshift-marketplace/community-operators-mvc44" Mar 17 10:17:26 crc kubenswrapper[4813]: I0317 10:17:26.644808 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6-utilities\") pod \"community-operators-mvc44\" (UID: \"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6\") " pod="openshift-marketplace/community-operators-mvc44" Mar 17 10:17:26 crc kubenswrapper[4813]: I0317 10:17:26.645425 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6-utilities\") pod \"community-operators-mvc44\" (UID: \"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6\") " pod="openshift-marketplace/community-operators-mvc44" Mar 17 10:17:26 crc kubenswrapper[4813]: I0317 10:17:26.645756 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6-catalog-content\") pod \"community-operators-mvc44\" (UID: \"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6\") " pod="openshift-marketplace/community-operators-mvc44" Mar 17 10:17:26 crc kubenswrapper[4813]: I0317 10:17:26.667484 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ql7zj\" (UniqueName: \"kubernetes.io/projected/b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6-kube-api-access-ql7zj\") pod \"community-operators-mvc44\" (UID: \"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6\") " pod="openshift-marketplace/community-operators-mvc44" Mar 17 10:17:26 crc kubenswrapper[4813]: I0317 10:17:26.695914 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mvc44" Mar 17 10:17:27 crc kubenswrapper[4813]: I0317 10:17:27.255087 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mvc44"] Mar 17 10:17:28 crc kubenswrapper[4813]: I0317 10:17:28.040114 4813 generic.go:334] "Generic (PLEG): container finished" podID="b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6" containerID="c85f6e5484947c78336444296ae20454b0aa7f09a53e4e6cd07dd5b3cb2f299d" exitCode=0 Mar 17 10:17:28 crc kubenswrapper[4813]: I0317 10:17:28.040173 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvc44" event={"ID":"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6","Type":"ContainerDied","Data":"c85f6e5484947c78336444296ae20454b0aa7f09a53e4e6cd07dd5b3cb2f299d"} Mar 17 10:17:28 crc kubenswrapper[4813]: I0317 10:17:28.040205 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvc44" event={"ID":"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6","Type":"ContainerStarted","Data":"395df92d5bafe468ee44d6154d93829dd4b4d674735e47c344e7e8d8178117b2"} Mar 17 10:17:29 crc kubenswrapper[4813]: I0317 10:17:29.050103 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvc44" event={"ID":"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6","Type":"ContainerStarted","Data":"7ea851b6fb260970c6dc460672753e36ef5a0b1edb364674a5e829f2fff3bb1c"} Mar 17 10:17:30 crc kubenswrapper[4813]: I0317 10:17:30.069122 4813 generic.go:334] "Generic (PLEG): container finished" podID="b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6" containerID="7ea851b6fb260970c6dc460672753e36ef5a0b1edb364674a5e829f2fff3bb1c" exitCode=0 Mar 17 10:17:30 crc kubenswrapper[4813]: I0317 10:17:30.069339 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvc44" event={"ID":"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6","Type":"ContainerDied","Data":"7ea851b6fb260970c6dc460672753e36ef5a0b1edb364674a5e829f2fff3bb1c"} Mar 17 10:17:31 crc kubenswrapper[4813]: I0317 10:17:31.083492 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvc44" event={"ID":"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6","Type":"ContainerStarted","Data":"f86a03fd1e2bbbd653426cf42d72a991f93f1725934b918b7efb3f3c1f6bc291"} Mar 17 10:17:31 crc kubenswrapper[4813]: I0317 10:17:31.104074 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mvc44" podStartSLOduration=2.6494676740000003 podStartE2EDuration="5.104053321s" podCreationTimestamp="2026-03-17 10:17:26 +0000 UTC" firstStartedPulling="2026-03-17 10:17:28.045306894 +0000 UTC m=+4070.146110403" lastFinishedPulling="2026-03-17 10:17:30.499892511 +0000 UTC m=+4072.600696050" observedRunningTime="2026-03-17 10:17:31.102955346 +0000 UTC m=+4073.203758855" watchObservedRunningTime="2026-03-17 10:17:31.104053321 +0000 UTC m=+4073.204856840" Mar 17 10:17:33 crc kubenswrapper[4813]: I0317 10:17:33.279104 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pzdlg/must-gather-sv27r"] Mar 17 10:17:33 crc kubenswrapper[4813]: I0317 10:17:33.282348 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pzdlg/must-gather-sv27r" Mar 17 10:17:33 crc kubenswrapper[4813]: I0317 10:17:33.285089 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-pzdlg"/"kube-root-ca.crt" Mar 17 10:17:33 crc kubenswrapper[4813]: I0317 10:17:33.285138 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-pzdlg"/"openshift-service-ca.crt" Mar 17 10:17:33 crc kubenswrapper[4813]: I0317 10:17:33.285508 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-pzdlg"/"default-dockercfg-9rfq6" Mar 17 10:17:33 crc kubenswrapper[4813]: I0317 10:17:33.292253 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-pzdlg/must-gather-sv27r"] Mar 17 10:17:33 crc kubenswrapper[4813]: I0317 10:17:33.396711 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzzmt\" (UniqueName: \"kubernetes.io/projected/397827d2-1331-46ef-8d9f-15f2ff101941-kube-api-access-tzzmt\") pod \"must-gather-sv27r\" (UID: \"397827d2-1331-46ef-8d9f-15f2ff101941\") " pod="openshift-must-gather-pzdlg/must-gather-sv27r" Mar 17 10:17:33 crc kubenswrapper[4813]: I0317 10:17:33.397085 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/397827d2-1331-46ef-8d9f-15f2ff101941-must-gather-output\") pod \"must-gather-sv27r\" (UID: \"397827d2-1331-46ef-8d9f-15f2ff101941\") " pod="openshift-must-gather-pzdlg/must-gather-sv27r" Mar 17 10:17:33 crc kubenswrapper[4813]: I0317 10:17:33.498306 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzzmt\" (UniqueName: \"kubernetes.io/projected/397827d2-1331-46ef-8d9f-15f2ff101941-kube-api-access-tzzmt\") pod \"must-gather-sv27r\" (UID: \"397827d2-1331-46ef-8d9f-15f2ff101941\") " pod="openshift-must-gather-pzdlg/must-gather-sv27r" Mar 17 10:17:33 crc kubenswrapper[4813]: I0317 10:17:33.498439 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/397827d2-1331-46ef-8d9f-15f2ff101941-must-gather-output\") pod \"must-gather-sv27r\" (UID: \"397827d2-1331-46ef-8d9f-15f2ff101941\") " pod="openshift-must-gather-pzdlg/must-gather-sv27r" Mar 17 10:17:33 crc kubenswrapper[4813]: I0317 10:17:33.498955 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/397827d2-1331-46ef-8d9f-15f2ff101941-must-gather-output\") pod \"must-gather-sv27r\" (UID: \"397827d2-1331-46ef-8d9f-15f2ff101941\") " pod="openshift-must-gather-pzdlg/must-gather-sv27r" Mar 17 10:17:33 crc kubenswrapper[4813]: I0317 10:17:33.518483 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzzmt\" (UniqueName: \"kubernetes.io/projected/397827d2-1331-46ef-8d9f-15f2ff101941-kube-api-access-tzzmt\") pod \"must-gather-sv27r\" (UID: \"397827d2-1331-46ef-8d9f-15f2ff101941\") " pod="openshift-must-gather-pzdlg/must-gather-sv27r" Mar 17 10:17:33 crc kubenswrapper[4813]: I0317 10:17:33.602537 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pzdlg/must-gather-sv27r" Mar 17 10:17:34 crc kubenswrapper[4813]: I0317 10:17:34.085119 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-pzdlg/must-gather-sv27r"] Mar 17 10:17:34 crc kubenswrapper[4813]: W0317 10:17:34.090109 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod397827d2_1331_46ef_8d9f_15f2ff101941.slice/crio-47bafa02c65ff8f853d9f4ecdcf7230a7b81d4b777f52b45e8306ff809fa8c9f WatchSource:0}: Error finding container 47bafa02c65ff8f853d9f4ecdcf7230a7b81d4b777f52b45e8306ff809fa8c9f: Status 404 returned error can't find the container with id 47bafa02c65ff8f853d9f4ecdcf7230a7b81d4b777f52b45e8306ff809fa8c9f Mar 17 10:17:34 crc kubenswrapper[4813]: I0317 10:17:34.112993 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pzdlg/must-gather-sv27r" event={"ID":"397827d2-1331-46ef-8d9f-15f2ff101941","Type":"ContainerStarted","Data":"47bafa02c65ff8f853d9f4ecdcf7230a7b81d4b777f52b45e8306ff809fa8c9f"} Mar 17 10:17:35 crc kubenswrapper[4813]: I0317 10:17:35.122515 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pzdlg/must-gather-sv27r" event={"ID":"397827d2-1331-46ef-8d9f-15f2ff101941","Type":"ContainerStarted","Data":"b6aaa4a49e47badfc0c2c7f28559dcdefbafad39fc8116a1fc85e0b45154ed02"} Mar 17 10:17:35 crc kubenswrapper[4813]: I0317 10:17:35.123733 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pzdlg/must-gather-sv27r" event={"ID":"397827d2-1331-46ef-8d9f-15f2ff101941","Type":"ContainerStarted","Data":"2d0f886accd63f5c18fd503b99df71bf14c6f1dffbcb8b80746726a1a296a767"} Mar 17 10:17:35 crc kubenswrapper[4813]: I0317 10:17:35.142749 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pzdlg/must-gather-sv27r" podStartSLOduration=2.142721878 podStartE2EDuration="2.142721878s" podCreationTimestamp="2026-03-17 10:17:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 10:17:35.138183178 +0000 UTC m=+4077.238986687" watchObservedRunningTime="2026-03-17 10:17:35.142721878 +0000 UTC m=+4077.243525417" Mar 17 10:17:36 crc kubenswrapper[4813]: I0317 10:17:36.697555 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mvc44" Mar 17 10:17:36 crc kubenswrapper[4813]: I0317 10:17:36.697985 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mvc44" Mar 17 10:17:36 crc kubenswrapper[4813]: I0317 10:17:36.781110 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mvc44" Mar 17 10:17:37 crc kubenswrapper[4813]: I0317 10:17:37.198390 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mvc44" Mar 17 10:17:37 crc kubenswrapper[4813]: I0317 10:17:37.248655 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mvc44"] Mar 17 10:17:37 crc kubenswrapper[4813]: I0317 10:17:37.831776 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pzdlg/crc-debug-q8w86"] Mar 17 10:17:37 crc kubenswrapper[4813]: I0317 10:17:37.833054 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pzdlg/crc-debug-q8w86" Mar 17 10:17:37 crc kubenswrapper[4813]: I0317 10:17:37.990954 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/555d71d8-9e5b-44ad-9508-087497e295e4-host\") pod \"crc-debug-q8w86\" (UID: \"555d71d8-9e5b-44ad-9508-087497e295e4\") " pod="openshift-must-gather-pzdlg/crc-debug-q8w86" Mar 17 10:17:37 crc kubenswrapper[4813]: I0317 10:17:37.991885 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxcq4\" (UniqueName: \"kubernetes.io/projected/555d71d8-9e5b-44ad-9508-087497e295e4-kube-api-access-qxcq4\") pod \"crc-debug-q8w86\" (UID: \"555d71d8-9e5b-44ad-9508-087497e295e4\") " pod="openshift-must-gather-pzdlg/crc-debug-q8w86" Mar 17 10:17:38 crc kubenswrapper[4813]: I0317 10:17:38.095409 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/555d71d8-9e5b-44ad-9508-087497e295e4-host\") pod \"crc-debug-q8w86\" (UID: \"555d71d8-9e5b-44ad-9508-087497e295e4\") " pod="openshift-must-gather-pzdlg/crc-debug-q8w86" Mar 17 10:17:38 crc kubenswrapper[4813]: I0317 10:17:38.095515 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxcq4\" (UniqueName: \"kubernetes.io/projected/555d71d8-9e5b-44ad-9508-087497e295e4-kube-api-access-qxcq4\") pod \"crc-debug-q8w86\" (UID: \"555d71d8-9e5b-44ad-9508-087497e295e4\") " pod="openshift-must-gather-pzdlg/crc-debug-q8w86" Mar 17 10:17:38 crc kubenswrapper[4813]: I0317 10:17:38.096029 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/555d71d8-9e5b-44ad-9508-087497e295e4-host\") pod \"crc-debug-q8w86\" (UID: \"555d71d8-9e5b-44ad-9508-087497e295e4\") " pod="openshift-must-gather-pzdlg/crc-debug-q8w86" Mar 17 10:17:38 crc kubenswrapper[4813]: I0317 10:17:38.125043 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxcq4\" (UniqueName: \"kubernetes.io/projected/555d71d8-9e5b-44ad-9508-087497e295e4-kube-api-access-qxcq4\") pod \"crc-debug-q8w86\" (UID: \"555d71d8-9e5b-44ad-9508-087497e295e4\") " pod="openshift-must-gather-pzdlg/crc-debug-q8w86" Mar 17 10:17:38 crc kubenswrapper[4813]: I0317 10:17:38.153044 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pzdlg/crc-debug-q8w86" Mar 17 10:17:39 crc kubenswrapper[4813]: I0317 10:17:39.154163 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mvc44" podUID="b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6" containerName="registry-server" containerID="cri-o://f86a03fd1e2bbbd653426cf42d72a991f93f1725934b918b7efb3f3c1f6bc291" gracePeriod=2 Mar 17 10:17:39 crc kubenswrapper[4813]: I0317 10:17:39.154922 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pzdlg/crc-debug-q8w86" event={"ID":"555d71d8-9e5b-44ad-9508-087497e295e4","Type":"ContainerStarted","Data":"1c5174094124ee6445542c7c1e2c617e783ffebef96e315ce1532dc37de6b0ce"} Mar 17 10:17:39 crc kubenswrapper[4813]: I0317 10:17:39.154953 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pzdlg/crc-debug-q8w86" event={"ID":"555d71d8-9e5b-44ad-9508-087497e295e4","Type":"ContainerStarted","Data":"927b2ad21688955c5e2a9d5ae0d05aae5cb579e222929001d36ddbc3de40b0cf"} Mar 17 10:17:39 crc kubenswrapper[4813]: I0317 10:17:39.188430 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pzdlg/crc-debug-q8w86" podStartSLOduration=2.188408855 podStartE2EDuration="2.188408855s" podCreationTimestamp="2026-03-17 10:17:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-17 10:17:39.173074528 +0000 UTC m=+4081.273878027" watchObservedRunningTime="2026-03-17 10:17:39.188408855 +0000 UTC m=+4081.289212354" Mar 17 10:17:39 crc kubenswrapper[4813]: I0317 10:17:39.878467 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mvc44" Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.042342 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6-catalog-content\") pod \"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6\" (UID: \"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6\") " Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.042523 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6-utilities\") pod \"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6\" (UID: \"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6\") " Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.042580 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ql7zj\" (UniqueName: \"kubernetes.io/projected/b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6-kube-api-access-ql7zj\") pod \"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6\" (UID: \"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6\") " Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.044241 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6-utilities" (OuterVolumeSpecName: "utilities") pod "b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6" (UID: "b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.059685 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6-kube-api-access-ql7zj" (OuterVolumeSpecName: "kube-api-access-ql7zj") pod "b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6" (UID: "b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6"). InnerVolumeSpecName "kube-api-access-ql7zj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.093029 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6" (UID: "b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.145091 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.145121 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ql7zj\" (UniqueName: \"kubernetes.io/projected/b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6-kube-api-access-ql7zj\") on node \"crc\" DevicePath \"\"" Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.145131 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.167113 4813 generic.go:334] "Generic (PLEG): container finished" podID="b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6" containerID="f86a03fd1e2bbbd653426cf42d72a991f93f1725934b918b7efb3f3c1f6bc291" exitCode=0 Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.167174 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvc44" event={"ID":"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6","Type":"ContainerDied","Data":"f86a03fd1e2bbbd653426cf42d72a991f93f1725934b918b7efb3f3c1f6bc291"} Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.167185 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mvc44" Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.167210 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mvc44" event={"ID":"b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6","Type":"ContainerDied","Data":"395df92d5bafe468ee44d6154d93829dd4b4d674735e47c344e7e8d8178117b2"} Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.167230 4813 scope.go:117] "RemoveContainer" containerID="f86a03fd1e2bbbd653426cf42d72a991f93f1725934b918b7efb3f3c1f6bc291" Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.192930 4813 scope.go:117] "RemoveContainer" containerID="7ea851b6fb260970c6dc460672753e36ef5a0b1edb364674a5e829f2fff3bb1c" Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.210368 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mvc44"] Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.217993 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mvc44"] Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.253534 4813 scope.go:117] "RemoveContainer" containerID="c85f6e5484947c78336444296ae20454b0aa7f09a53e4e6cd07dd5b3cb2f299d" Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.281328 4813 scope.go:117] "RemoveContainer" containerID="f86a03fd1e2bbbd653426cf42d72a991f93f1725934b918b7efb3f3c1f6bc291" Mar 17 10:17:40 crc kubenswrapper[4813]: E0317 10:17:40.283381 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f86a03fd1e2bbbd653426cf42d72a991f93f1725934b918b7efb3f3c1f6bc291\": container with ID starting with f86a03fd1e2bbbd653426cf42d72a991f93f1725934b918b7efb3f3c1f6bc291 not found: ID does not exist" containerID="f86a03fd1e2bbbd653426cf42d72a991f93f1725934b918b7efb3f3c1f6bc291" Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.283421 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f86a03fd1e2bbbd653426cf42d72a991f93f1725934b918b7efb3f3c1f6bc291"} err="failed to get container status \"f86a03fd1e2bbbd653426cf42d72a991f93f1725934b918b7efb3f3c1f6bc291\": rpc error: code = NotFound desc = could not find container \"f86a03fd1e2bbbd653426cf42d72a991f93f1725934b918b7efb3f3c1f6bc291\": container with ID starting with f86a03fd1e2bbbd653426cf42d72a991f93f1725934b918b7efb3f3c1f6bc291 not found: ID does not exist" Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.283448 4813 scope.go:117] "RemoveContainer" containerID="7ea851b6fb260970c6dc460672753e36ef5a0b1edb364674a5e829f2fff3bb1c" Mar 17 10:17:40 crc kubenswrapper[4813]: E0317 10:17:40.290844 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ea851b6fb260970c6dc460672753e36ef5a0b1edb364674a5e829f2fff3bb1c\": container with ID starting with 7ea851b6fb260970c6dc460672753e36ef5a0b1edb364674a5e829f2fff3bb1c not found: ID does not exist" containerID="7ea851b6fb260970c6dc460672753e36ef5a0b1edb364674a5e829f2fff3bb1c" Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.290887 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ea851b6fb260970c6dc460672753e36ef5a0b1edb364674a5e829f2fff3bb1c"} err="failed to get container status \"7ea851b6fb260970c6dc460672753e36ef5a0b1edb364674a5e829f2fff3bb1c\": rpc error: code = NotFound desc = could not find container \"7ea851b6fb260970c6dc460672753e36ef5a0b1edb364674a5e829f2fff3bb1c\": container with ID starting with 7ea851b6fb260970c6dc460672753e36ef5a0b1edb364674a5e829f2fff3bb1c not found: ID does not exist" Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.290914 4813 scope.go:117] "RemoveContainer" containerID="c85f6e5484947c78336444296ae20454b0aa7f09a53e4e6cd07dd5b3cb2f299d" Mar 17 10:17:40 crc kubenswrapper[4813]: E0317 10:17:40.291389 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c85f6e5484947c78336444296ae20454b0aa7f09a53e4e6cd07dd5b3cb2f299d\": container with ID starting with c85f6e5484947c78336444296ae20454b0aa7f09a53e4e6cd07dd5b3cb2f299d not found: ID does not exist" containerID="c85f6e5484947c78336444296ae20454b0aa7f09a53e4e6cd07dd5b3cb2f299d" Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.291435 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c85f6e5484947c78336444296ae20454b0aa7f09a53e4e6cd07dd5b3cb2f299d"} err="failed to get container status \"c85f6e5484947c78336444296ae20454b0aa7f09a53e4e6cd07dd5b3cb2f299d\": rpc error: code = NotFound desc = could not find container \"c85f6e5484947c78336444296ae20454b0aa7f09a53e4e6cd07dd5b3cb2f299d\": container with ID starting with c85f6e5484947c78336444296ae20454b0aa7f09a53e4e6cd07dd5b3cb2f299d not found: ID does not exist" Mar 17 10:17:40 crc kubenswrapper[4813]: I0317 10:17:40.740739 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6" path="/var/lib/kubelet/pods/b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6/volumes" Mar 17 10:18:00 crc kubenswrapper[4813]: I0317 10:18:00.153344 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562378-4ghd7"] Mar 17 10:18:00 crc kubenswrapper[4813]: E0317 10:18:00.154285 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6" containerName="extract-content" Mar 17 10:18:00 crc kubenswrapper[4813]: I0317 10:18:00.154300 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6" containerName="extract-content" Mar 17 10:18:00 crc kubenswrapper[4813]: E0317 10:18:00.154319 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6" containerName="extract-utilities" Mar 17 10:18:00 crc kubenswrapper[4813]: I0317 10:18:00.154327 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6" containerName="extract-utilities" Mar 17 10:18:00 crc kubenswrapper[4813]: E0317 10:18:00.154345 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6" containerName="registry-server" Mar 17 10:18:00 crc kubenswrapper[4813]: I0317 10:18:00.154352 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6" containerName="registry-server" Mar 17 10:18:00 crc kubenswrapper[4813]: I0317 10:18:00.154546 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b21355d7-2f8d-4b2e-b6af-c3edb86e3cf6" containerName="registry-server" Mar 17 10:18:00 crc kubenswrapper[4813]: I0317 10:18:00.155248 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562378-4ghd7" Mar 17 10:18:00 crc kubenswrapper[4813]: I0317 10:18:00.159789 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 10:18:00 crc kubenswrapper[4813]: I0317 10:18:00.159976 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 10:18:00 crc kubenswrapper[4813]: I0317 10:18:00.160077 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 10:18:00 crc kubenswrapper[4813]: I0317 10:18:00.168908 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562378-4ghd7"] Mar 17 10:18:00 crc kubenswrapper[4813]: I0317 10:18:00.262880 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rw6pn\" (UniqueName: \"kubernetes.io/projected/e34845ff-00c2-44f9-a3ed-fb16e52e1a06-kube-api-access-rw6pn\") pod \"auto-csr-approver-29562378-4ghd7\" (UID: \"e34845ff-00c2-44f9-a3ed-fb16e52e1a06\") " pod="openshift-infra/auto-csr-approver-29562378-4ghd7" Mar 17 10:18:00 crc kubenswrapper[4813]: I0317 10:18:00.364544 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rw6pn\" (UniqueName: \"kubernetes.io/projected/e34845ff-00c2-44f9-a3ed-fb16e52e1a06-kube-api-access-rw6pn\") pod \"auto-csr-approver-29562378-4ghd7\" (UID: \"e34845ff-00c2-44f9-a3ed-fb16e52e1a06\") " pod="openshift-infra/auto-csr-approver-29562378-4ghd7" Mar 17 10:18:00 crc kubenswrapper[4813]: I0317 10:18:00.385523 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rw6pn\" (UniqueName: \"kubernetes.io/projected/e34845ff-00c2-44f9-a3ed-fb16e52e1a06-kube-api-access-rw6pn\") pod \"auto-csr-approver-29562378-4ghd7\" (UID: \"e34845ff-00c2-44f9-a3ed-fb16e52e1a06\") " pod="openshift-infra/auto-csr-approver-29562378-4ghd7" Mar 17 10:18:00 crc kubenswrapper[4813]: I0317 10:18:00.481369 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562378-4ghd7" Mar 17 10:18:00 crc kubenswrapper[4813]: I0317 10:18:00.971546 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562378-4ghd7"] Mar 17 10:18:01 crc kubenswrapper[4813]: I0317 10:18:01.343216 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562378-4ghd7" event={"ID":"e34845ff-00c2-44f9-a3ed-fb16e52e1a06","Type":"ContainerStarted","Data":"46ee9c572cc1313a5f89f89f1cd17a91fb96d3915bc6dc645afd98ce9952fda7"} Mar 17 10:18:03 crc kubenswrapper[4813]: I0317 10:18:03.367516 4813 generic.go:334] "Generic (PLEG): container finished" podID="e34845ff-00c2-44f9-a3ed-fb16e52e1a06" containerID="cc3051458887669a97f7638605718a52e19ac68b4bff84fa90ef107aa3d50d8e" exitCode=0 Mar 17 10:18:03 crc kubenswrapper[4813]: I0317 10:18:03.367849 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562378-4ghd7" event={"ID":"e34845ff-00c2-44f9-a3ed-fb16e52e1a06","Type":"ContainerDied","Data":"cc3051458887669a97f7638605718a52e19ac68b4bff84fa90ef107aa3d50d8e"} Mar 17 10:18:04 crc kubenswrapper[4813]: I0317 10:18:04.796401 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562378-4ghd7" Mar 17 10:18:04 crc kubenswrapper[4813]: I0317 10:18:04.950175 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rw6pn\" (UniqueName: \"kubernetes.io/projected/e34845ff-00c2-44f9-a3ed-fb16e52e1a06-kube-api-access-rw6pn\") pod \"e34845ff-00c2-44f9-a3ed-fb16e52e1a06\" (UID: \"e34845ff-00c2-44f9-a3ed-fb16e52e1a06\") " Mar 17 10:18:04 crc kubenswrapper[4813]: I0317 10:18:04.956879 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e34845ff-00c2-44f9-a3ed-fb16e52e1a06-kube-api-access-rw6pn" (OuterVolumeSpecName: "kube-api-access-rw6pn") pod "e34845ff-00c2-44f9-a3ed-fb16e52e1a06" (UID: "e34845ff-00c2-44f9-a3ed-fb16e52e1a06"). InnerVolumeSpecName "kube-api-access-rw6pn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:18:05 crc kubenswrapper[4813]: I0317 10:18:05.053240 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rw6pn\" (UniqueName: \"kubernetes.io/projected/e34845ff-00c2-44f9-a3ed-fb16e52e1a06-kube-api-access-rw6pn\") on node \"crc\" DevicePath \"\"" Mar 17 10:18:05 crc kubenswrapper[4813]: I0317 10:18:05.384069 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562378-4ghd7" event={"ID":"e34845ff-00c2-44f9-a3ed-fb16e52e1a06","Type":"ContainerDied","Data":"46ee9c572cc1313a5f89f89f1cd17a91fb96d3915bc6dc645afd98ce9952fda7"} Mar 17 10:18:05 crc kubenswrapper[4813]: I0317 10:18:05.384102 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46ee9c572cc1313a5f89f89f1cd17a91fb96d3915bc6dc645afd98ce9952fda7" Mar 17 10:18:05 crc kubenswrapper[4813]: I0317 10:18:05.384506 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562378-4ghd7" Mar 17 10:18:05 crc kubenswrapper[4813]: I0317 10:18:05.867872 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562372-jrqr7"] Mar 17 10:18:05 crc kubenswrapper[4813]: I0317 10:18:05.876665 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562372-jrqr7"] Mar 17 10:18:06 crc kubenswrapper[4813]: I0317 10:18:06.748096 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7b6e6f5-528c-40fa-8f42-892a57e71fee" path="/var/lib/kubelet/pods/b7b6e6f5-528c-40fa-8f42-892a57e71fee/volumes" Mar 17 10:18:12 crc kubenswrapper[4813]: I0317 10:18:12.458536 4813 generic.go:334] "Generic (PLEG): container finished" podID="555d71d8-9e5b-44ad-9508-087497e295e4" containerID="1c5174094124ee6445542c7c1e2c617e783ffebef96e315ce1532dc37de6b0ce" exitCode=0 Mar 17 10:18:12 crc kubenswrapper[4813]: I0317 10:18:12.458592 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pzdlg/crc-debug-q8w86" event={"ID":"555d71d8-9e5b-44ad-9508-087497e295e4","Type":"ContainerDied","Data":"1c5174094124ee6445542c7c1e2c617e783ffebef96e315ce1532dc37de6b0ce"} Mar 17 10:18:13 crc kubenswrapper[4813]: I0317 10:18:13.568141 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pzdlg/crc-debug-q8w86" Mar 17 10:18:13 crc kubenswrapper[4813]: I0317 10:18:13.607454 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pzdlg/crc-debug-q8w86"] Mar 17 10:18:13 crc kubenswrapper[4813]: I0317 10:18:13.617771 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pzdlg/crc-debug-q8w86"] Mar 17 10:18:13 crc kubenswrapper[4813]: I0317 10:18:13.714741 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/555d71d8-9e5b-44ad-9508-087497e295e4-host\") pod \"555d71d8-9e5b-44ad-9508-087497e295e4\" (UID: \"555d71d8-9e5b-44ad-9508-087497e295e4\") " Mar 17 10:18:13 crc kubenswrapper[4813]: I0317 10:18:13.714862 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/555d71d8-9e5b-44ad-9508-087497e295e4-host" (OuterVolumeSpecName: "host") pod "555d71d8-9e5b-44ad-9508-087497e295e4" (UID: "555d71d8-9e5b-44ad-9508-087497e295e4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 10:18:13 crc kubenswrapper[4813]: I0317 10:18:13.714905 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qxcq4\" (UniqueName: \"kubernetes.io/projected/555d71d8-9e5b-44ad-9508-087497e295e4-kube-api-access-qxcq4\") pod \"555d71d8-9e5b-44ad-9508-087497e295e4\" (UID: \"555d71d8-9e5b-44ad-9508-087497e295e4\") " Mar 17 10:18:13 crc kubenswrapper[4813]: I0317 10:18:13.715410 4813 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/555d71d8-9e5b-44ad-9508-087497e295e4-host\") on node \"crc\" DevicePath \"\"" Mar 17 10:18:13 crc kubenswrapper[4813]: I0317 10:18:13.720951 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/555d71d8-9e5b-44ad-9508-087497e295e4-kube-api-access-qxcq4" (OuterVolumeSpecName: "kube-api-access-qxcq4") pod "555d71d8-9e5b-44ad-9508-087497e295e4" (UID: "555d71d8-9e5b-44ad-9508-087497e295e4"). InnerVolumeSpecName "kube-api-access-qxcq4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:18:13 crc kubenswrapper[4813]: I0317 10:18:13.817096 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qxcq4\" (UniqueName: \"kubernetes.io/projected/555d71d8-9e5b-44ad-9508-087497e295e4-kube-api-access-qxcq4\") on node \"crc\" DevicePath \"\"" Mar 17 10:18:14 crc kubenswrapper[4813]: I0317 10:18:14.479245 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="927b2ad21688955c5e2a9d5ae0d05aae5cb579e222929001d36ddbc3de40b0cf" Mar 17 10:18:14 crc kubenswrapper[4813]: I0317 10:18:14.479311 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pzdlg/crc-debug-q8w86" Mar 17 10:18:14 crc kubenswrapper[4813]: I0317 10:18:14.740755 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="555d71d8-9e5b-44ad-9508-087497e295e4" path="/var/lib/kubelet/pods/555d71d8-9e5b-44ad-9508-087497e295e4/volumes" Mar 17 10:18:14 crc kubenswrapper[4813]: I0317 10:18:14.825253 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pzdlg/crc-debug-cz4dt"] Mar 17 10:18:14 crc kubenswrapper[4813]: E0317 10:18:14.825634 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e34845ff-00c2-44f9-a3ed-fb16e52e1a06" containerName="oc" Mar 17 10:18:14 crc kubenswrapper[4813]: I0317 10:18:14.825649 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="e34845ff-00c2-44f9-a3ed-fb16e52e1a06" containerName="oc" Mar 17 10:18:14 crc kubenswrapper[4813]: E0317 10:18:14.825681 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="555d71d8-9e5b-44ad-9508-087497e295e4" containerName="container-00" Mar 17 10:18:14 crc kubenswrapper[4813]: I0317 10:18:14.825687 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="555d71d8-9e5b-44ad-9508-087497e295e4" containerName="container-00" Mar 17 10:18:14 crc kubenswrapper[4813]: I0317 10:18:14.825846 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="e34845ff-00c2-44f9-a3ed-fb16e52e1a06" containerName="oc" Mar 17 10:18:14 crc kubenswrapper[4813]: I0317 10:18:14.825861 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="555d71d8-9e5b-44ad-9508-087497e295e4" containerName="container-00" Mar 17 10:18:14 crc kubenswrapper[4813]: I0317 10:18:14.826423 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pzdlg/crc-debug-cz4dt" Mar 17 10:18:14 crc kubenswrapper[4813]: I0317 10:18:14.835469 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90d467e6-004c-4a21-b714-3ea72ded4acd-host\") pod \"crc-debug-cz4dt\" (UID: \"90d467e6-004c-4a21-b714-3ea72ded4acd\") " pod="openshift-must-gather-pzdlg/crc-debug-cz4dt" Mar 17 10:18:14 crc kubenswrapper[4813]: I0317 10:18:14.835756 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xzfcl\" (UniqueName: \"kubernetes.io/projected/90d467e6-004c-4a21-b714-3ea72ded4acd-kube-api-access-xzfcl\") pod \"crc-debug-cz4dt\" (UID: \"90d467e6-004c-4a21-b714-3ea72ded4acd\") " pod="openshift-must-gather-pzdlg/crc-debug-cz4dt" Mar 17 10:18:14 crc kubenswrapper[4813]: I0317 10:18:14.937293 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xzfcl\" (UniqueName: \"kubernetes.io/projected/90d467e6-004c-4a21-b714-3ea72ded4acd-kube-api-access-xzfcl\") pod \"crc-debug-cz4dt\" (UID: \"90d467e6-004c-4a21-b714-3ea72ded4acd\") " pod="openshift-must-gather-pzdlg/crc-debug-cz4dt" Mar 17 10:18:14 crc kubenswrapper[4813]: I0317 10:18:14.937392 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90d467e6-004c-4a21-b714-3ea72ded4acd-host\") pod \"crc-debug-cz4dt\" (UID: \"90d467e6-004c-4a21-b714-3ea72ded4acd\") " pod="openshift-must-gather-pzdlg/crc-debug-cz4dt" Mar 17 10:18:14 crc kubenswrapper[4813]: I0317 10:18:14.937528 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90d467e6-004c-4a21-b714-3ea72ded4acd-host\") pod \"crc-debug-cz4dt\" (UID: \"90d467e6-004c-4a21-b714-3ea72ded4acd\") " pod="openshift-must-gather-pzdlg/crc-debug-cz4dt" Mar 17 10:18:14 crc kubenswrapper[4813]: I0317 10:18:14.960090 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xzfcl\" (UniqueName: \"kubernetes.io/projected/90d467e6-004c-4a21-b714-3ea72ded4acd-kube-api-access-xzfcl\") pod \"crc-debug-cz4dt\" (UID: \"90d467e6-004c-4a21-b714-3ea72ded4acd\") " pod="openshift-must-gather-pzdlg/crc-debug-cz4dt" Mar 17 10:18:15 crc kubenswrapper[4813]: I0317 10:18:15.141398 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pzdlg/crc-debug-cz4dt" Mar 17 10:18:15 crc kubenswrapper[4813]: W0317 10:18:15.180477 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90d467e6_004c_4a21_b714_3ea72ded4acd.slice/crio-bb46bc07eff54e7d8a61ceebc03a91edef4843c54b87f1d143d97b71b0d68e71 WatchSource:0}: Error finding container bb46bc07eff54e7d8a61ceebc03a91edef4843c54b87f1d143d97b71b0d68e71: Status 404 returned error can't find the container with id bb46bc07eff54e7d8a61ceebc03a91edef4843c54b87f1d143d97b71b0d68e71 Mar 17 10:18:15 crc kubenswrapper[4813]: I0317 10:18:15.490782 4813 generic.go:334] "Generic (PLEG): container finished" podID="90d467e6-004c-4a21-b714-3ea72ded4acd" containerID="67b5aa7d74f11399737349586b688846b3c4e4424b2e9090a021c764f8c25251" exitCode=0 Mar 17 10:18:15 crc kubenswrapper[4813]: I0317 10:18:15.490849 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pzdlg/crc-debug-cz4dt" event={"ID":"90d467e6-004c-4a21-b714-3ea72ded4acd","Type":"ContainerDied","Data":"67b5aa7d74f11399737349586b688846b3c4e4424b2e9090a021c764f8c25251"} Mar 17 10:18:15 crc kubenswrapper[4813]: I0317 10:18:15.490884 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pzdlg/crc-debug-cz4dt" event={"ID":"90d467e6-004c-4a21-b714-3ea72ded4acd","Type":"ContainerStarted","Data":"bb46bc07eff54e7d8a61ceebc03a91edef4843c54b87f1d143d97b71b0d68e71"} Mar 17 10:18:15 crc kubenswrapper[4813]: I0317 10:18:15.968667 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pzdlg/crc-debug-cz4dt"] Mar 17 10:18:15 crc kubenswrapper[4813]: I0317 10:18:15.975721 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pzdlg/crc-debug-cz4dt"] Mar 17 10:18:16 crc kubenswrapper[4813]: I0317 10:18:16.595166 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pzdlg/crc-debug-cz4dt" Mar 17 10:18:16 crc kubenswrapper[4813]: I0317 10:18:16.772319 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90d467e6-004c-4a21-b714-3ea72ded4acd-host\") pod \"90d467e6-004c-4a21-b714-3ea72ded4acd\" (UID: \"90d467e6-004c-4a21-b714-3ea72ded4acd\") " Mar 17 10:18:16 crc kubenswrapper[4813]: I0317 10:18:16.772392 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xzfcl\" (UniqueName: \"kubernetes.io/projected/90d467e6-004c-4a21-b714-3ea72ded4acd-kube-api-access-xzfcl\") pod \"90d467e6-004c-4a21-b714-3ea72ded4acd\" (UID: \"90d467e6-004c-4a21-b714-3ea72ded4acd\") " Mar 17 10:18:16 crc kubenswrapper[4813]: I0317 10:18:16.772423 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/90d467e6-004c-4a21-b714-3ea72ded4acd-host" (OuterVolumeSpecName: "host") pod "90d467e6-004c-4a21-b714-3ea72ded4acd" (UID: "90d467e6-004c-4a21-b714-3ea72ded4acd"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 10:18:16 crc kubenswrapper[4813]: I0317 10:18:16.774243 4813 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/90d467e6-004c-4a21-b714-3ea72ded4acd-host\") on node \"crc\" DevicePath \"\"" Mar 17 10:18:16 crc kubenswrapper[4813]: I0317 10:18:16.778430 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/90d467e6-004c-4a21-b714-3ea72ded4acd-kube-api-access-xzfcl" (OuterVolumeSpecName: "kube-api-access-xzfcl") pod "90d467e6-004c-4a21-b714-3ea72ded4acd" (UID: "90d467e6-004c-4a21-b714-3ea72ded4acd"). InnerVolumeSpecName "kube-api-access-xzfcl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:18:16 crc kubenswrapper[4813]: I0317 10:18:16.876280 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xzfcl\" (UniqueName: \"kubernetes.io/projected/90d467e6-004c-4a21-b714-3ea72ded4acd-kube-api-access-xzfcl\") on node \"crc\" DevicePath \"\"" Mar 17 10:18:17 crc kubenswrapper[4813]: I0317 10:18:17.233075 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pzdlg/crc-debug-cnct9"] Mar 17 10:18:17 crc kubenswrapper[4813]: E0317 10:18:17.233760 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="90d467e6-004c-4a21-b714-3ea72ded4acd" containerName="container-00" Mar 17 10:18:17 crc kubenswrapper[4813]: I0317 10:18:17.233772 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="90d467e6-004c-4a21-b714-3ea72ded4acd" containerName="container-00" Mar 17 10:18:17 crc kubenswrapper[4813]: I0317 10:18:17.233958 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="90d467e6-004c-4a21-b714-3ea72ded4acd" containerName="container-00" Mar 17 10:18:17 crc kubenswrapper[4813]: I0317 10:18:17.234559 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pzdlg/crc-debug-cnct9" Mar 17 10:18:17 crc kubenswrapper[4813]: I0317 10:18:17.281106 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b394c104-1f81-47ad-9bf1-e45a2376d725-host\") pod \"crc-debug-cnct9\" (UID: \"b394c104-1f81-47ad-9bf1-e45a2376d725\") " pod="openshift-must-gather-pzdlg/crc-debug-cnct9" Mar 17 10:18:17 crc kubenswrapper[4813]: I0317 10:18:17.281201 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v68sj\" (UniqueName: \"kubernetes.io/projected/b394c104-1f81-47ad-9bf1-e45a2376d725-kube-api-access-v68sj\") pod \"crc-debug-cnct9\" (UID: \"b394c104-1f81-47ad-9bf1-e45a2376d725\") " pod="openshift-must-gather-pzdlg/crc-debug-cnct9" Mar 17 10:18:17 crc kubenswrapper[4813]: I0317 10:18:17.383067 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b394c104-1f81-47ad-9bf1-e45a2376d725-host\") pod \"crc-debug-cnct9\" (UID: \"b394c104-1f81-47ad-9bf1-e45a2376d725\") " pod="openshift-must-gather-pzdlg/crc-debug-cnct9" Mar 17 10:18:17 crc kubenswrapper[4813]: I0317 10:18:17.383185 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b394c104-1f81-47ad-9bf1-e45a2376d725-host\") pod \"crc-debug-cnct9\" (UID: \"b394c104-1f81-47ad-9bf1-e45a2376d725\") " pod="openshift-must-gather-pzdlg/crc-debug-cnct9" Mar 17 10:18:17 crc kubenswrapper[4813]: I0317 10:18:17.383195 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v68sj\" (UniqueName: \"kubernetes.io/projected/b394c104-1f81-47ad-9bf1-e45a2376d725-kube-api-access-v68sj\") pod \"crc-debug-cnct9\" (UID: \"b394c104-1f81-47ad-9bf1-e45a2376d725\") " pod="openshift-must-gather-pzdlg/crc-debug-cnct9" Mar 17 10:18:17 crc kubenswrapper[4813]: I0317 10:18:17.406196 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v68sj\" (UniqueName: \"kubernetes.io/projected/b394c104-1f81-47ad-9bf1-e45a2376d725-kube-api-access-v68sj\") pod \"crc-debug-cnct9\" (UID: \"b394c104-1f81-47ad-9bf1-e45a2376d725\") " pod="openshift-must-gather-pzdlg/crc-debug-cnct9" Mar 17 10:18:17 crc kubenswrapper[4813]: I0317 10:18:17.511207 4813 scope.go:117] "RemoveContainer" containerID="67b5aa7d74f11399737349586b688846b3c4e4424b2e9090a021c764f8c25251" Mar 17 10:18:17 crc kubenswrapper[4813]: I0317 10:18:17.511330 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pzdlg/crc-debug-cz4dt" Mar 17 10:18:17 crc kubenswrapper[4813]: I0317 10:18:17.555085 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pzdlg/crc-debug-cnct9" Mar 17 10:18:17 crc kubenswrapper[4813]: W0317 10:18:17.583883 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb394c104_1f81_47ad_9bf1_e45a2376d725.slice/crio-d5bd4e7fea46fa054ce5b2a6fae6e5434ad9a2ed6b362d8f8a4c71513d7e8896 WatchSource:0}: Error finding container d5bd4e7fea46fa054ce5b2a6fae6e5434ad9a2ed6b362d8f8a4c71513d7e8896: Status 404 returned error can't find the container with id d5bd4e7fea46fa054ce5b2a6fae6e5434ad9a2ed6b362d8f8a4c71513d7e8896 Mar 17 10:18:18 crc kubenswrapper[4813]: I0317 10:18:18.521298 4813 generic.go:334] "Generic (PLEG): container finished" podID="b394c104-1f81-47ad-9bf1-e45a2376d725" containerID="168743feeacae9cbd663ea8a71346f98ecf5c0e5ead1a00c8b36b2b9d5f484c7" exitCode=0 Mar 17 10:18:18 crc kubenswrapper[4813]: I0317 10:18:18.521463 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pzdlg/crc-debug-cnct9" event={"ID":"b394c104-1f81-47ad-9bf1-e45a2376d725","Type":"ContainerDied","Data":"168743feeacae9cbd663ea8a71346f98ecf5c0e5ead1a00c8b36b2b9d5f484c7"} Mar 17 10:18:18 crc kubenswrapper[4813]: I0317 10:18:18.521584 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pzdlg/crc-debug-cnct9" event={"ID":"b394c104-1f81-47ad-9bf1-e45a2376d725","Type":"ContainerStarted","Data":"d5bd4e7fea46fa054ce5b2a6fae6e5434ad9a2ed6b362d8f8a4c71513d7e8896"} Mar 17 10:18:18 crc kubenswrapper[4813]: I0317 10:18:18.568989 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pzdlg/crc-debug-cnct9"] Mar 17 10:18:18 crc kubenswrapper[4813]: I0317 10:18:18.581049 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pzdlg/crc-debug-cnct9"] Mar 17 10:18:18 crc kubenswrapper[4813]: I0317 10:18:18.747507 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="90d467e6-004c-4a21-b714-3ea72ded4acd" path="/var/lib/kubelet/pods/90d467e6-004c-4a21-b714-3ea72ded4acd/volumes" Mar 17 10:18:19 crc kubenswrapper[4813]: I0317 10:18:19.619956 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pzdlg/crc-debug-cnct9" Mar 17 10:18:19 crc kubenswrapper[4813]: I0317 10:18:19.633417 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v68sj\" (UniqueName: \"kubernetes.io/projected/b394c104-1f81-47ad-9bf1-e45a2376d725-kube-api-access-v68sj\") pod \"b394c104-1f81-47ad-9bf1-e45a2376d725\" (UID: \"b394c104-1f81-47ad-9bf1-e45a2376d725\") " Mar 17 10:18:19 crc kubenswrapper[4813]: I0317 10:18:19.633561 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b394c104-1f81-47ad-9bf1-e45a2376d725-host\") pod \"b394c104-1f81-47ad-9bf1-e45a2376d725\" (UID: \"b394c104-1f81-47ad-9bf1-e45a2376d725\") " Mar 17 10:18:19 crc kubenswrapper[4813]: I0317 10:18:19.634092 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b394c104-1f81-47ad-9bf1-e45a2376d725-host" (OuterVolumeSpecName: "host") pod "b394c104-1f81-47ad-9bf1-e45a2376d725" (UID: "b394c104-1f81-47ad-9bf1-e45a2376d725"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 17 10:18:19 crc kubenswrapper[4813]: I0317 10:18:19.641768 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b394c104-1f81-47ad-9bf1-e45a2376d725-kube-api-access-v68sj" (OuterVolumeSpecName: "kube-api-access-v68sj") pod "b394c104-1f81-47ad-9bf1-e45a2376d725" (UID: "b394c104-1f81-47ad-9bf1-e45a2376d725"). InnerVolumeSpecName "kube-api-access-v68sj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:18:19 crc kubenswrapper[4813]: I0317 10:18:19.735295 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v68sj\" (UniqueName: \"kubernetes.io/projected/b394c104-1f81-47ad-9bf1-e45a2376d725-kube-api-access-v68sj\") on node \"crc\" DevicePath \"\"" Mar 17 10:18:19 crc kubenswrapper[4813]: I0317 10:18:19.735338 4813 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b394c104-1f81-47ad-9bf1-e45a2376d725-host\") on node \"crc\" DevicePath \"\"" Mar 17 10:18:20 crc kubenswrapper[4813]: I0317 10:18:20.542976 4813 scope.go:117] "RemoveContainer" containerID="168743feeacae9cbd663ea8a71346f98ecf5c0e5ead1a00c8b36b2b9d5f484c7" Mar 17 10:18:20 crc kubenswrapper[4813]: I0317 10:18:20.543126 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pzdlg/crc-debug-cnct9" Mar 17 10:18:20 crc kubenswrapper[4813]: I0317 10:18:20.740185 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b394c104-1f81-47ad-9bf1-e45a2376d725" path="/var/lib/kubelet/pods/b394c104-1f81-47ad-9bf1-e45a2376d725/volumes" Mar 17 10:18:44 crc kubenswrapper[4813]: I0317 10:18:44.114005 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 10:18:44 crc kubenswrapper[4813]: I0317 10:18:44.114575 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 10:18:51 crc kubenswrapper[4813]: I0317 10:18:51.503878 4813 scope.go:117] "RemoveContainer" containerID="71f20f0035a4216c4aba6e5b93bc999c58c77eb58e565bc3704fca0d4432bccc" Mar 17 10:18:51 crc kubenswrapper[4813]: I0317 10:18:51.946566 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-79bdd687db-2rnkr_80441abc-02b7-4a55-b4b9-2c3c7adf8ed2/barbican-api/0.log" Mar 17 10:18:52 crc kubenswrapper[4813]: I0317 10:18:52.074004 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-79bdd687db-2rnkr_80441abc-02b7-4a55-b4b9-2c3c7adf8ed2/barbican-api-log/0.log" Mar 17 10:18:52 crc kubenswrapper[4813]: I0317 10:18:52.187211 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5ff5576dbb-hlqck_2d44bf48-4229-453c-925d-9653d547de9e/barbican-keystone-listener/0.log" Mar 17 10:18:52 crc kubenswrapper[4813]: I0317 10:18:52.235571 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-5ff5576dbb-hlqck_2d44bf48-4229-453c-925d-9653d547de9e/barbican-keystone-listener-log/0.log" Mar 17 10:18:52 crc kubenswrapper[4813]: I0317 10:18:52.326767 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-76556857b7-95bv6_91f4bba0-7f6d-43cf-8887-00b081856e89/barbican-worker/0.log" Mar 17 10:18:52 crc kubenswrapper[4813]: I0317 10:18:52.381991 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-76556857b7-95bv6_91f4bba0-7f6d-43cf-8887-00b081856e89/barbican-worker-log/0.log" Mar 17 10:18:52 crc kubenswrapper[4813]: I0317 10:18:52.514785 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-c9plg_f9166d03-0282-47e6-a80b-5b03f4183f62/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:18:52 crc kubenswrapper[4813]: I0317 10:18:52.603491 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ea9cbc63-ff93-4342-ab4c-7ddf7bf78577/ceilometer-central-agent/0.log" Mar 17 10:18:52 crc kubenswrapper[4813]: I0317 10:18:52.704164 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ea9cbc63-ff93-4342-ab4c-7ddf7bf78577/proxy-httpd/0.log" Mar 17 10:18:52 crc kubenswrapper[4813]: I0317 10:18:52.733772 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ea9cbc63-ff93-4342-ab4c-7ddf7bf78577/ceilometer-notification-agent/0.log" Mar 17 10:18:52 crc kubenswrapper[4813]: I0317 10:18:52.745490 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_ea9cbc63-ff93-4342-ab4c-7ddf7bf78577/sg-core/0.log" Mar 17 10:18:52 crc kubenswrapper[4813]: I0317 10:18:52.907766 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_27d30ca5-0f9f-4f22-83a6-66c9e0942930/cinder-api-log/0.log" Mar 17 10:18:52 crc kubenswrapper[4813]: I0317 10:18:52.916432 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_27d30ca5-0f9f-4f22-83a6-66c9e0942930/cinder-api/0.log" Mar 17 10:18:53 crc kubenswrapper[4813]: I0317 10:18:53.104261 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_e2cad367-9c0f-4556-9e4c-b6173aa1b161/cinder-scheduler/0.log" Mar 17 10:18:53 crc kubenswrapper[4813]: I0317 10:18:53.145807 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_e2cad367-9c0f-4556-9e4c-b6173aa1b161/probe/0.log" Mar 17 10:18:53 crc kubenswrapper[4813]: I0317 10:18:53.576129 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-4kzll_c7908d9c-b027-4b03-8879-37968eea2f28/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:18:53 crc kubenswrapper[4813]: I0317 10:18:53.608674 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-vt5m5_0d8211b8-b1c4-4ec3-8dcd-5b7beefc7fdc/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:18:53 crc kubenswrapper[4813]: I0317 10:18:53.747501 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-drzd4_3cceef96-6dbd-4173-b363-3bc4fbf4f598/init/0.log" Mar 17 10:18:53 crc kubenswrapper[4813]: I0317 10:18:53.941241 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-drzd4_3cceef96-6dbd-4173-b363-3bc4fbf4f598/init/0.log" Mar 17 10:18:53 crc kubenswrapper[4813]: I0317 10:18:53.992671 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cb6ffcf87-drzd4_3cceef96-6dbd-4173-b363-3bc4fbf4f598/dnsmasq-dns/0.log" Mar 17 10:18:54 crc kubenswrapper[4813]: I0317 10:18:54.051867 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_download-cache-edpm-deployment-openstack-edpm-ipam-46np8_36536921-03a1-45dd-93b3-1d06e9c3adca/download-cache-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:18:54 crc kubenswrapper[4813]: I0317 10:18:54.246696 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f4b4ddd1-d517-43cc-ac5e-636026efff5d/glance-httpd/0.log" Mar 17 10:18:54 crc kubenswrapper[4813]: I0317 10:18:54.322156 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_f4b4ddd1-d517-43cc-ac5e-636026efff5d/glance-log/0.log" Mar 17 10:18:54 crc kubenswrapper[4813]: I0317 10:18:54.413666 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7056da71-3a67-4258-8f12-9ab7b50a83ea/glance-httpd/0.log" Mar 17 10:18:54 crc kubenswrapper[4813]: I0317 10:18:54.467072 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_7056da71-3a67-4258-8f12-9ab7b50a83ea/glance-log/0.log" Mar 17 10:18:54 crc kubenswrapper[4813]: I0317 10:18:54.600859 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5dc999f6f8-t94dl_bf22966e-516d-40c2-975c-c3e41122b8d2/horizon/0.log" Mar 17 10:18:54 crc kubenswrapper[4813]: I0317 10:18:54.799830 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-pjjzt_d8cecca4-b584-4537-810f-02c1818d2e99/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:18:54 crc kubenswrapper[4813]: I0317 10:18:54.937216 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-64lbt_63ad7386-248c-4dc8-a217-fb2cf3d4ef82/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:18:54 crc kubenswrapper[4813]: I0317 10:18:54.979876 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5dc999f6f8-t94dl_bf22966e-516d-40c2-975c-c3e41122b8d2/horizon-log/0.log" Mar 17 10:18:55 crc kubenswrapper[4813]: I0317 10:18:55.179791 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-7fb6796ddc-58j2l_fd0582a1-002f-452c-828a-406a4d945f28/keystone-api/0.log" Mar 17 10:18:55 crc kubenswrapper[4813]: I0317 10:18:55.185273 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29562361-p6pcq_8a4ef497-31ad-415a-8c23-1673c0e279cb/keystone-cron/0.log" Mar 17 10:18:55 crc kubenswrapper[4813]: I0317 10:18:55.497552 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_9e83f1d1-1126-49a5-9a24-1694e7616b61/kube-state-metrics/0.log" Mar 17 10:18:55 crc kubenswrapper[4813]: I0317 10:18:55.529133 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-2gfq8_71e0074f-5ee2-4f59-9184-c23495021bfd/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:18:55 crc kubenswrapper[4813]: I0317 10:18:55.888305 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-76bd9cc6f7-x2p6w_7d7e90bc-8bc2-4c06-9131-405fa21670df/neutron-api/0.log" Mar 17 10:18:55 crc kubenswrapper[4813]: I0317 10:18:55.936350 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-76bd9cc6f7-x2p6w_7d7e90bc-8bc2-4c06-9131-405fa21670df/neutron-httpd/0.log" Mar 17 10:18:56 crc kubenswrapper[4813]: I0317 10:18:56.128907 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-7fbpj_0fb2d682-7ad5-4c0f-bbf3-5fceb1461eab/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:18:56 crc kubenswrapper[4813]: I0317 10:18:56.535268 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_04ea2c77-25c5-45fc-a375-cf04b185381c/nova-api-log/0.log" Mar 17 10:18:56 crc kubenswrapper[4813]: I0317 10:18:56.605613 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_08fc5042-2349-4425-a338-e9f9ab5cc125/nova-cell0-conductor-conductor/0.log" Mar 17 10:18:57 crc kubenswrapper[4813]: I0317 10:18:57.015022 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_c8cfa16e-a939-4e80-9e09-2a632e7bd29e/nova-cell1-conductor-conductor/0.log" Mar 17 10:18:57 crc kubenswrapper[4813]: I0317 10:18:57.060042 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_9cc3a926-a367-45b1-9d38-6f673720e71e/nova-cell1-novncproxy-novncproxy/0.log" Mar 17 10:18:57 crc kubenswrapper[4813]: I0317 10:18:57.189174 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_04ea2c77-25c5-45fc-a375-cf04b185381c/nova-api-api/0.log" Mar 17 10:18:57 crc kubenswrapper[4813]: I0317 10:18:57.202636 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-edpm-deployment-openstack-edpm-ipam-6z5zk_bd00eea2-860c-40ac-8147-41992af388ec/nova-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:18:57 crc kubenswrapper[4813]: I0317 10:18:57.337039 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_82bd221d-6ac0-4eb2-b709-3de76c656745/nova-metadata-log/0.log" Mar 17 10:18:57 crc kubenswrapper[4813]: I0317 10:18:57.663666 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1/mysql-bootstrap/0.log" Mar 17 10:18:57 crc kubenswrapper[4813]: I0317 10:18:57.696920 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_aa41f41c-74ec-44a3-a913-afd1a44d4d04/nova-scheduler-scheduler/0.log" Mar 17 10:18:57 crc kubenswrapper[4813]: I0317 10:18:57.798484 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_82bd221d-6ac0-4eb2-b709-3de76c656745/nova-metadata-metadata/0.log" Mar 17 10:18:57 crc kubenswrapper[4813]: I0317 10:18:57.838774 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1/galera/0.log" Mar 17 10:18:57 crc kubenswrapper[4813]: I0317 10:18:57.872516 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_95fc0ebb-17fc-49b0-b8a5-a6c3ed7d4bd1/mysql-bootstrap/0.log" Mar 17 10:18:58 crc kubenswrapper[4813]: I0317 10:18:58.027418 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9ad18f78-70c7-4b1d-bda7-ed338fdcdd64/mysql-bootstrap/0.log" Mar 17 10:18:58 crc kubenswrapper[4813]: I0317 10:18:58.261731 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9ad18f78-70c7-4b1d-bda7-ed338fdcdd64/mysql-bootstrap/0.log" Mar 17 10:18:58 crc kubenswrapper[4813]: I0317 10:18:58.286068 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_9ad18f78-70c7-4b1d-bda7-ed338fdcdd64/galera/0.log" Mar 17 10:18:58 crc kubenswrapper[4813]: I0317 10:18:58.333371 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_c532fc6b-8310-4322-b09c-bd21b19154ec/openstackclient/0.log" Mar 17 10:18:58 crc kubenswrapper[4813]: I0317 10:18:58.488098 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-btlxg_8cbfe69c-a2c5-40bf-a8d2-e78755f060f2/openstack-network-exporter/0.log" Mar 17 10:18:58 crc kubenswrapper[4813]: I0317 10:18:58.619741 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bt47r_3752b62f-49dd-46a7-ae34-9b6b20343a09/ovsdb-server-init/0.log" Mar 17 10:18:58 crc kubenswrapper[4813]: I0317 10:18:58.809538 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bt47r_3752b62f-49dd-46a7-ae34-9b6b20343a09/ovsdb-server-init/0.log" Mar 17 10:18:58 crc kubenswrapper[4813]: I0317 10:18:58.873879 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bt47r_3752b62f-49dd-46a7-ae34-9b6b20343a09/ovsdb-server/0.log" Mar 17 10:18:58 crc kubenswrapper[4813]: I0317 10:18:58.916261 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-bt47r_3752b62f-49dd-46a7-ae34-9b6b20343a09/ovs-vswitchd/0.log" Mar 17 10:18:59 crc kubenswrapper[4813]: I0317 10:18:59.055177 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-vdc6c_4090993b-51ce-4ce3-a6d6-a1501ab3ba05/ovn-controller/0.log" Mar 17 10:18:59 crc kubenswrapper[4813]: I0317 10:18:59.148791 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-c5ttr_5e4d8ae8-8ff6-4969-a349-813dde113094/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:18:59 crc kubenswrapper[4813]: I0317 10:18:59.407159 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_b433a82f-1948-4fdc-a395-86a0c07fee36/openstack-network-exporter/0.log" Mar 17 10:18:59 crc kubenswrapper[4813]: I0317 10:18:59.489315 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_b433a82f-1948-4fdc-a395-86a0c07fee36/ovn-northd/0.log" Mar 17 10:18:59 crc kubenswrapper[4813]: I0317 10:18:59.625908 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_34e286fb-7dd1-4e85-89e3-e926b232f5a6/ovsdbserver-nb/0.log" Mar 17 10:18:59 crc kubenswrapper[4813]: I0317 10:18:59.693383 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_34e286fb-7dd1-4e85-89e3-e926b232f5a6/openstack-network-exporter/0.log" Mar 17 10:18:59 crc kubenswrapper[4813]: I0317 10:18:59.799410 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7e4c8a33-829e-4aec-be7c-0dad92ce3916/openstack-network-exporter/0.log" Mar 17 10:18:59 crc kubenswrapper[4813]: I0317 10:18:59.848709 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_7e4c8a33-829e-4aec-be7c-0dad92ce3916/ovsdbserver-sb/0.log" Mar 17 10:19:00 crc kubenswrapper[4813]: I0317 10:19:00.407039 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-557566c676-bbf8q_46eacf89-88ff-4be7-a8c1-e90784324da2/placement-api/0.log" Mar 17 10:19:00 crc kubenswrapper[4813]: I0317 10:19:00.442843 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cfc379d0-d53a-4640-8d3f-a883db1aefa9/setup-container/0.log" Mar 17 10:19:00 crc kubenswrapper[4813]: I0317 10:19:00.528157 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-557566c676-bbf8q_46eacf89-88ff-4be7-a8c1-e90784324da2/placement-log/0.log" Mar 17 10:19:00 crc kubenswrapper[4813]: I0317 10:19:00.597514 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cfc379d0-d53a-4640-8d3f-a883db1aefa9/setup-container/0.log" Mar 17 10:19:00 crc kubenswrapper[4813]: I0317 10:19:00.619004 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_cfc379d0-d53a-4640-8d3f-a883db1aefa9/rabbitmq/0.log" Mar 17 10:19:00 crc kubenswrapper[4813]: I0317 10:19:00.764357 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b/setup-container/0.log" Mar 17 10:19:00 crc kubenswrapper[4813]: I0317 10:19:00.969105 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b/setup-container/0.log" Mar 17 10:19:00 crc kubenswrapper[4813]: I0317 10:19:00.990110 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_16d776ec-b4ca-4cf2-8ab7-1e73bc6a137b/rabbitmq/0.log" Mar 17 10:19:01 crc kubenswrapper[4813]: I0317 10:19:01.072991 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-shjmk_60339fa3-6329-4aa7-a958-d28be7f562cc/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:19:01 crc kubenswrapper[4813]: I0317 10:19:01.212792 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_redhat-edpm-deployment-openstack-edpm-ipam-5blh4_c0a603db-d69a-47a6-8228-862e4ad835ee/redhat-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:19:01 crc kubenswrapper[4813]: I0317 10:19:01.336908 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-pdzdr_df99907b-2dfb-436f-b9a0-5c62dce82672/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:19:01 crc kubenswrapper[4813]: I0317 10:19:01.412290 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-jhl6m_540d96f7-a4a1-490f-9c57-997a2c6dc9ab/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:19:01 crc kubenswrapper[4813]: I0317 10:19:01.979907 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-ht9q9_1d6a2187-6711-4b3f-988a-845c5b7e18f7/ssh-known-hosts-edpm-deployment/0.log" Mar 17 10:19:02 crc kubenswrapper[4813]: I0317 10:19:02.119463 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-8557f6579f-shvsw_e79b20b0-682e-4d16-bb56-64f0c4ec0202/proxy-httpd/0.log" Mar 17 10:19:02 crc kubenswrapper[4813]: I0317 10:19:02.230356 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-8557f6579f-shvsw_e79b20b0-682e-4d16-bb56-64f0c4ec0202/proxy-server/0.log" Mar 17 10:19:02 crc kubenswrapper[4813]: I0317 10:19:02.350463 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-rrfrc_625a6e48-f069-49a9-b11a-342e12dffece/swift-ring-rebalance/0.log" Mar 17 10:19:02 crc kubenswrapper[4813]: I0317 10:19:02.509259 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/account-auditor/0.log" Mar 17 10:19:02 crc kubenswrapper[4813]: I0317 10:19:02.514289 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/account-reaper/0.log" Mar 17 10:19:02 crc kubenswrapper[4813]: I0317 10:19:02.621587 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/account-server/0.log" Mar 17 10:19:02 crc kubenswrapper[4813]: I0317 10:19:02.626614 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/account-replicator/0.log" Mar 17 10:19:02 crc kubenswrapper[4813]: I0317 10:19:02.721520 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/container-auditor/0.log" Mar 17 10:19:02 crc kubenswrapper[4813]: I0317 10:19:02.748074 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/container-replicator/0.log" Mar 17 10:19:02 crc kubenswrapper[4813]: I0317 10:19:02.822868 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/container-server/0.log" Mar 17 10:19:02 crc kubenswrapper[4813]: I0317 10:19:02.869981 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/container-updater/0.log" Mar 17 10:19:03 crc kubenswrapper[4813]: I0317 10:19:03.009533 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/object-auditor/0.log" Mar 17 10:19:03 crc kubenswrapper[4813]: I0317 10:19:03.040414 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/object-expirer/0.log" Mar 17 10:19:03 crc kubenswrapper[4813]: I0317 10:19:03.052025 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/object-replicator/0.log" Mar 17 10:19:03 crc kubenswrapper[4813]: I0317 10:19:03.300297 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/object-server/0.log" Mar 17 10:19:03 crc kubenswrapper[4813]: I0317 10:19:03.343375 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/object-updater/0.log" Mar 17 10:19:03 crc kubenswrapper[4813]: I0317 10:19:03.406508 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/rsync/0.log" Mar 17 10:19:03 crc kubenswrapper[4813]: I0317 10:19:03.417304 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_76ba87f0-f2dc-4099-a8d9-7ece7c3e7c51/swift-recon-cron/0.log" Mar 17 10:19:03 crc kubenswrapper[4813]: I0317 10:19:03.608901 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-972kp_cd1a58de-c5c8-4b15-a894-47baac35d6e2/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:19:03 crc kubenswrapper[4813]: I0317 10:19:03.651781 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_005f04b0-01f5-4bed-9d2c-2269ab67d27f/tempest-tests-tempest-tests-runner/0.log" Mar 17 10:19:03 crc kubenswrapper[4813]: I0317 10:19:03.857812 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_5b58dbf4-d51d-4fe8-90d0-d295217084e1/test-operator-logs-container/0.log" Mar 17 10:19:03 crc kubenswrapper[4813]: I0317 10:19:03.915364 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-vs6qm_0c2d5562-76da-433b-9840-59385eeb872c/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Mar 17 10:19:12 crc kubenswrapper[4813]: I0317 10:19:12.124456 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_c43912b9-949e-4d64-ae70-e59594cc329a/memcached/0.log" Mar 17 10:19:14 crc kubenswrapper[4813]: I0317 10:19:14.114007 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 10:19:14 crc kubenswrapper[4813]: I0317 10:19:14.114079 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 10:19:31 crc kubenswrapper[4813]: I0317 10:19:31.884814 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd_fa9ca0da-b53a-42bc-ace9-49579d21ead8/util/0.log" Mar 17 10:19:32 crc kubenswrapper[4813]: I0317 10:19:32.042172 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd_fa9ca0da-b53a-42bc-ace9-49579d21ead8/pull/0.log" Mar 17 10:19:32 crc kubenswrapper[4813]: I0317 10:19:32.056389 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd_fa9ca0da-b53a-42bc-ace9-49579d21ead8/util/0.log" Mar 17 10:19:32 crc kubenswrapper[4813]: I0317 10:19:32.074805 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd_fa9ca0da-b53a-42bc-ace9-49579d21ead8/pull/0.log" Mar 17 10:19:32 crc kubenswrapper[4813]: I0317 10:19:32.211214 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd_fa9ca0da-b53a-42bc-ace9-49579d21ead8/util/0.log" Mar 17 10:19:32 crc kubenswrapper[4813]: I0317 10:19:32.223346 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd_fa9ca0da-b53a-42bc-ace9-49579d21ead8/pull/0.log" Mar 17 10:19:32 crc kubenswrapper[4813]: I0317 10:19:32.228625 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_47ef5cb47a81e259808e9647fbb85eb0871323e2c224f97995d5dd96a5959xd_fa9ca0da-b53a-42bc-ace9-49579d21ead8/extract/0.log" Mar 17 10:19:32 crc kubenswrapper[4813]: I0317 10:19:32.439589 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-59bc569d95-xkfkm_227f29cf-ec7d-40e9-b28c-c1ccae9d880e/manager/0.log" Mar 17 10:19:32 crc kubenswrapper[4813]: I0317 10:19:32.727863 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-588d4d986b-tn4pv_9b3cf477-8c47-4a38-beee-8bd3f70977e0/manager/0.log" Mar 17 10:19:32 crc kubenswrapper[4813]: I0317 10:19:32.923056 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-67dd5f86f5-hng2s_de6c9679-9f9f-481d-85d9-218c9199088e/manager/0.log" Mar 17 10:19:32 crc kubenswrapper[4813]: I0317 10:19:32.925076 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-79df6bcc97-wzmxx_3c5a1075-8105-401b-afb0-00014dafed0a/manager/0.log" Mar 17 10:19:33 crc kubenswrapper[4813]: I0317 10:19:33.139783 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-8464cc45fb-tfhsc_668691cc-39f3-4b04-8ed6-bc86f9965f12/manager/0.log" Mar 17 10:19:33 crc kubenswrapper[4813]: I0317 10:19:33.455068 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6f787dddc9-4ds8c_80f14d5d-c8b9-4068-9469-ea8eff96e9e5/manager/0.log" Mar 17 10:19:33 crc kubenswrapper[4813]: I0317 10:19:33.664936 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-7b9c774f96-7965b_c9ea4c59-d87b-4e58-ac8f-2786f8f32f02/manager/0.log" Mar 17 10:19:33 crc kubenswrapper[4813]: I0317 10:19:33.733167 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-768b96df4c-c966n_fd3396be-4825-4edc-a41f-443af1413e3b/manager/0.log" Mar 17 10:19:33 crc kubenswrapper[4813]: I0317 10:19:33.803784 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-8d58dc466-9cqth_1b2b8088-31e5-49af-91b4-a3fbfe797af5/manager/0.log" Mar 17 10:19:33 crc kubenswrapper[4813]: I0317 10:19:33.858919 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-55f864c847-dbs68_5e7b137d-9b58-448b-91a4-3a069dfb4d10/manager/0.log" Mar 17 10:19:34 crc kubenswrapper[4813]: I0317 10:19:34.000942 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-67ccfc9778-ngd7n_0e3eba88-a4a4-4c6d-9e90-5dba41268494/manager/0.log" Mar 17 10:19:34 crc kubenswrapper[4813]: I0317 10:19:34.083214 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-767865f676-mnsb4_c3419efc-507c-46bc-aedb-ef5eb3f0f917/manager/0.log" Mar 17 10:19:34 crc kubenswrapper[4813]: I0317 10:19:34.252049 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-5d488d59fb-tqg57_b4b9a61e-819f-4da5-a28a-8e49c0a3f7cd/manager/0.log" Mar 17 10:19:34 crc kubenswrapper[4813]: I0317 10:19:34.287023 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-5b9f45d989-mps6t_c6e1a616-6d48-4ef1-93cb-62caed2de963/manager/0.log" Mar 17 10:19:34 crc kubenswrapper[4813]: I0317 10:19:34.442586 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-89d64c458-g5vgd_8e80dcf7-dae9-4faa-ab3f-b547b570926b/manager/0.log" Mar 17 10:19:34 crc kubenswrapper[4813]: I0317 10:19:34.576910 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-init-56ddf7fcb8-7hkdj_6dc9eebf-39dd-4030-a091-5626832fd141/operator/0.log" Mar 17 10:19:34 crc kubenswrapper[4813]: I0317 10:19:34.896033 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-z8wcr_e56922bf-05f4-474c-b883-f1ca16e3e1cd/registry-server/0.log" Mar 17 10:19:35 crc kubenswrapper[4813]: I0317 10:19:35.041816 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-884679f54-7d66d_76eb47fe-8f18-4f98-a8ba-d89169d0de93/manager/0.log" Mar 17 10:19:35 crc kubenswrapper[4813]: I0317 10:19:35.225246 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-5784578c99-hcsvd_968158c3-2dfc-439e-9d20-8706d611fec1/manager/0.log" Mar 17 10:19:35 crc kubenswrapper[4813]: I0317 10:19:35.286037 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-rxzjs_bcd2954f-b240-4cc3-8387-15b0e64a9721/operator/0.log" Mar 17 10:19:35 crc kubenswrapper[4813]: I0317 10:19:35.450117 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-c674c5965-njxx7_ac052585-f733-47d5-91b9-87ad7957e511/manager/0.log" Mar 17 10:19:35 crc kubenswrapper[4813]: I0317 10:19:35.590776 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-d6b694c5-52xx2_ec48dbf8-730b-4a2e-9986-03d8f66de013/manager/0.log" Mar 17 10:19:35 crc kubenswrapper[4813]: I0317 10:19:35.658345 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5c5cb9c4d7-r2ksk_860ccadc-ca76-4523-9543-2b0960eb77a9/manager/0.log" Mar 17 10:19:35 crc kubenswrapper[4813]: I0317 10:19:35.791991 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-65f8859ffd-mxfzh_db193bcc-748f-4830-873c-48a2f9229b40/manager/0.log" Mar 17 10:19:35 crc kubenswrapper[4813]: I0317 10:19:35.816546 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-6c4d75f7f9-wrtl9_520d3ab3-78d7-4be3-8466-9dadfce15c2c/manager/0.log" Mar 17 10:19:44 crc kubenswrapper[4813]: I0317 10:19:44.114162 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 10:19:44 crc kubenswrapper[4813]: I0317 10:19:44.114885 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 10:19:44 crc kubenswrapper[4813]: I0317 10:19:44.114951 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 10:19:44 crc kubenswrapper[4813]: I0317 10:19:44.116105 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"38fdc8059c00804eb8afc6b3d9dac72a950b31fe31c73eb825628373cb5b059d"} pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 10:19:44 crc kubenswrapper[4813]: I0317 10:19:44.116204 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" containerID="cri-o://38fdc8059c00804eb8afc6b3d9dac72a950b31fe31c73eb825628373cb5b059d" gracePeriod=600 Mar 17 10:19:44 crc kubenswrapper[4813]: I0317 10:19:44.505946 4813 generic.go:334] "Generic (PLEG): container finished" podID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerID="38fdc8059c00804eb8afc6b3d9dac72a950b31fe31c73eb825628373cb5b059d" exitCode=0 Mar 17 10:19:44 crc kubenswrapper[4813]: I0317 10:19:44.506074 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerDied","Data":"38fdc8059c00804eb8afc6b3d9dac72a950b31fe31c73eb825628373cb5b059d"} Mar 17 10:19:44 crc kubenswrapper[4813]: I0317 10:19:44.506312 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerStarted","Data":"c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee"} Mar 17 10:19:44 crc kubenswrapper[4813]: I0317 10:19:44.506342 4813 scope.go:117] "RemoveContainer" containerID="bfe36d606fe266aebadc91dab401f324433dffbc80b5ce6b69d3c0eb603c65f9" Mar 17 10:19:58 crc kubenswrapper[4813]: I0317 10:19:58.250117 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-sk47b_feae3098-e93c-4dff-9088-a38db6a5d929/kube-rbac-proxy/0.log" Mar 17 10:19:58 crc kubenswrapper[4813]: I0317 10:19:58.277722 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-j5djq_37f3773a-2367-411b-b8e4-c4ae96970de3/control-plane-machine-set-operator/0.log" Mar 17 10:19:58 crc kubenswrapper[4813]: I0317 10:19:58.396060 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-sk47b_feae3098-e93c-4dff-9088-a38db6a5d929/machine-api-operator/0.log" Mar 17 10:20:00 crc kubenswrapper[4813]: I0317 10:20:00.178648 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562380-p499h"] Mar 17 10:20:00 crc kubenswrapper[4813]: E0317 10:20:00.179296 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b394c104-1f81-47ad-9bf1-e45a2376d725" containerName="container-00" Mar 17 10:20:00 crc kubenswrapper[4813]: I0317 10:20:00.179307 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="b394c104-1f81-47ad-9bf1-e45a2376d725" containerName="container-00" Mar 17 10:20:00 crc kubenswrapper[4813]: I0317 10:20:00.179472 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="b394c104-1f81-47ad-9bf1-e45a2376d725" containerName="container-00" Mar 17 10:20:00 crc kubenswrapper[4813]: I0317 10:20:00.180092 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562380-p499h" Mar 17 10:20:00 crc kubenswrapper[4813]: I0317 10:20:00.182068 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 10:20:00 crc kubenswrapper[4813]: I0317 10:20:00.183322 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 10:20:00 crc kubenswrapper[4813]: I0317 10:20:00.187205 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 10:20:00 crc kubenswrapper[4813]: I0317 10:20:00.194362 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562380-p499h"] Mar 17 10:20:00 crc kubenswrapper[4813]: I0317 10:20:00.289342 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45k8d\" (UniqueName: \"kubernetes.io/projected/bf51d593-228d-4052-aa60-54849843ee9a-kube-api-access-45k8d\") pod \"auto-csr-approver-29562380-p499h\" (UID: \"bf51d593-228d-4052-aa60-54849843ee9a\") " pod="openshift-infra/auto-csr-approver-29562380-p499h" Mar 17 10:20:00 crc kubenswrapper[4813]: I0317 10:20:00.391111 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45k8d\" (UniqueName: \"kubernetes.io/projected/bf51d593-228d-4052-aa60-54849843ee9a-kube-api-access-45k8d\") pod \"auto-csr-approver-29562380-p499h\" (UID: \"bf51d593-228d-4052-aa60-54849843ee9a\") " pod="openshift-infra/auto-csr-approver-29562380-p499h" Mar 17 10:20:00 crc kubenswrapper[4813]: I0317 10:20:00.418789 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45k8d\" (UniqueName: \"kubernetes.io/projected/bf51d593-228d-4052-aa60-54849843ee9a-kube-api-access-45k8d\") pod \"auto-csr-approver-29562380-p499h\" (UID: \"bf51d593-228d-4052-aa60-54849843ee9a\") " pod="openshift-infra/auto-csr-approver-29562380-p499h" Mar 17 10:20:00 crc kubenswrapper[4813]: I0317 10:20:00.500346 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562380-p499h" Mar 17 10:20:00 crc kubenswrapper[4813]: I0317 10:20:00.983184 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562380-p499h"] Mar 17 10:20:00 crc kubenswrapper[4813]: W0317 10:20:00.988613 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbf51d593_228d_4052_aa60_54849843ee9a.slice/crio-44c6a41fd1c8a8f9d3e217c6cc07baaeb3e1569fdd6dda2e3519736a1b768135 WatchSource:0}: Error finding container 44c6a41fd1c8a8f9d3e217c6cc07baaeb3e1569fdd6dda2e3519736a1b768135: Status 404 returned error can't find the container with id 44c6a41fd1c8a8f9d3e217c6cc07baaeb3e1569fdd6dda2e3519736a1b768135 Mar 17 10:20:01 crc kubenswrapper[4813]: I0317 10:20:01.696716 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562380-p499h" event={"ID":"bf51d593-228d-4052-aa60-54849843ee9a","Type":"ContainerStarted","Data":"44c6a41fd1c8a8f9d3e217c6cc07baaeb3e1569fdd6dda2e3519736a1b768135"} Mar 17 10:20:02 crc kubenswrapper[4813]: I0317 10:20:02.712925 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562380-p499h" event={"ID":"bf51d593-228d-4052-aa60-54849843ee9a","Type":"ContainerStarted","Data":"0661df999c2af64696b83686113e55692c266b6d6108c7b5de72bb526c3d38cb"} Mar 17 10:20:02 crc kubenswrapper[4813]: I0317 10:20:02.730713 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562380-p499h" podStartSLOduration=1.41964366 podStartE2EDuration="2.730687272s" podCreationTimestamp="2026-03-17 10:20:00 +0000 UTC" firstStartedPulling="2026-03-17 10:20:00.996249939 +0000 UTC m=+4223.097053428" lastFinishedPulling="2026-03-17 10:20:02.307293501 +0000 UTC m=+4224.408097040" observedRunningTime="2026-03-17 10:20:02.726847033 +0000 UTC m=+4224.827650562" watchObservedRunningTime="2026-03-17 10:20:02.730687272 +0000 UTC m=+4224.831490821" Mar 17 10:20:03 crc kubenswrapper[4813]: I0317 10:20:03.725633 4813 generic.go:334] "Generic (PLEG): container finished" podID="bf51d593-228d-4052-aa60-54849843ee9a" containerID="0661df999c2af64696b83686113e55692c266b6d6108c7b5de72bb526c3d38cb" exitCode=0 Mar 17 10:20:03 crc kubenswrapper[4813]: I0317 10:20:03.726035 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562380-p499h" event={"ID":"bf51d593-228d-4052-aa60-54849843ee9a","Type":"ContainerDied","Data":"0661df999c2af64696b83686113e55692c266b6d6108c7b5de72bb526c3d38cb"} Mar 17 10:20:05 crc kubenswrapper[4813]: I0317 10:20:05.093302 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562380-p499h" Mar 17 10:20:05 crc kubenswrapper[4813]: I0317 10:20:05.189173 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-45k8d\" (UniqueName: \"kubernetes.io/projected/bf51d593-228d-4052-aa60-54849843ee9a-kube-api-access-45k8d\") pod \"bf51d593-228d-4052-aa60-54849843ee9a\" (UID: \"bf51d593-228d-4052-aa60-54849843ee9a\") " Mar 17 10:20:05 crc kubenswrapper[4813]: I0317 10:20:05.194516 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf51d593-228d-4052-aa60-54849843ee9a-kube-api-access-45k8d" (OuterVolumeSpecName: "kube-api-access-45k8d") pod "bf51d593-228d-4052-aa60-54849843ee9a" (UID: "bf51d593-228d-4052-aa60-54849843ee9a"). InnerVolumeSpecName "kube-api-access-45k8d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:20:05 crc kubenswrapper[4813]: I0317 10:20:05.291822 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-45k8d\" (UniqueName: \"kubernetes.io/projected/bf51d593-228d-4052-aa60-54849843ee9a-kube-api-access-45k8d\") on node \"crc\" DevicePath \"\"" Mar 17 10:20:05 crc kubenswrapper[4813]: I0317 10:20:05.759070 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562380-p499h" event={"ID":"bf51d593-228d-4052-aa60-54849843ee9a","Type":"ContainerDied","Data":"44c6a41fd1c8a8f9d3e217c6cc07baaeb3e1569fdd6dda2e3519736a1b768135"} Mar 17 10:20:05 crc kubenswrapper[4813]: I0317 10:20:05.759171 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="44c6a41fd1c8a8f9d3e217c6cc07baaeb3e1569fdd6dda2e3519736a1b768135" Mar 17 10:20:05 crc kubenswrapper[4813]: I0317 10:20:05.759205 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562380-p499h" Mar 17 10:20:05 crc kubenswrapper[4813]: I0317 10:20:05.811849 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562374-sx7jn"] Mar 17 10:20:05 crc kubenswrapper[4813]: I0317 10:20:05.819983 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562374-sx7jn"] Mar 17 10:20:06 crc kubenswrapper[4813]: I0317 10:20:06.749825 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c4330a08-3146-4c6d-aa2e-ba94a5d7d24c" path="/var/lib/kubelet/pods/c4330a08-3146-4c6d-aa2e-ba94a5d7d24c/volumes" Mar 17 10:20:09 crc kubenswrapper[4813]: I0317 10:20:09.115232 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5658z"] Mar 17 10:20:09 crc kubenswrapper[4813]: E0317 10:20:09.116252 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bf51d593-228d-4052-aa60-54849843ee9a" containerName="oc" Mar 17 10:20:09 crc kubenswrapper[4813]: I0317 10:20:09.116278 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="bf51d593-228d-4052-aa60-54849843ee9a" containerName="oc" Mar 17 10:20:09 crc kubenswrapper[4813]: I0317 10:20:09.116634 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="bf51d593-228d-4052-aa60-54849843ee9a" containerName="oc" Mar 17 10:20:09 crc kubenswrapper[4813]: I0317 10:20:09.118962 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5658z" Mar 17 10:20:09 crc kubenswrapper[4813]: I0317 10:20:09.125781 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5658z"] Mar 17 10:20:09 crc kubenswrapper[4813]: I0317 10:20:09.284255 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08aaed36-4ec7-4a1e-9a9b-fc78eb29726b-utilities\") pod \"redhat-marketplace-5658z\" (UID: \"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b\") " pod="openshift-marketplace/redhat-marketplace-5658z" Mar 17 10:20:09 crc kubenswrapper[4813]: I0317 10:20:09.284357 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08aaed36-4ec7-4a1e-9a9b-fc78eb29726b-catalog-content\") pod \"redhat-marketplace-5658z\" (UID: \"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b\") " pod="openshift-marketplace/redhat-marketplace-5658z" Mar 17 10:20:09 crc kubenswrapper[4813]: I0317 10:20:09.284673 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxswz\" (UniqueName: \"kubernetes.io/projected/08aaed36-4ec7-4a1e-9a9b-fc78eb29726b-kube-api-access-gxswz\") pod \"redhat-marketplace-5658z\" (UID: \"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b\") " pod="openshift-marketplace/redhat-marketplace-5658z" Mar 17 10:20:09 crc kubenswrapper[4813]: I0317 10:20:09.386744 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08aaed36-4ec7-4a1e-9a9b-fc78eb29726b-utilities\") pod \"redhat-marketplace-5658z\" (UID: \"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b\") " pod="openshift-marketplace/redhat-marketplace-5658z" Mar 17 10:20:09 crc kubenswrapper[4813]: I0317 10:20:09.386838 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08aaed36-4ec7-4a1e-9a9b-fc78eb29726b-catalog-content\") pod \"redhat-marketplace-5658z\" (UID: \"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b\") " pod="openshift-marketplace/redhat-marketplace-5658z" Mar 17 10:20:09 crc kubenswrapper[4813]: I0317 10:20:09.386950 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxswz\" (UniqueName: \"kubernetes.io/projected/08aaed36-4ec7-4a1e-9a9b-fc78eb29726b-kube-api-access-gxswz\") pod \"redhat-marketplace-5658z\" (UID: \"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b\") " pod="openshift-marketplace/redhat-marketplace-5658z" Mar 17 10:20:09 crc kubenswrapper[4813]: I0317 10:20:09.387406 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08aaed36-4ec7-4a1e-9a9b-fc78eb29726b-utilities\") pod \"redhat-marketplace-5658z\" (UID: \"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b\") " pod="openshift-marketplace/redhat-marketplace-5658z" Mar 17 10:20:09 crc kubenswrapper[4813]: I0317 10:20:09.387434 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08aaed36-4ec7-4a1e-9a9b-fc78eb29726b-catalog-content\") pod \"redhat-marketplace-5658z\" (UID: \"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b\") " pod="openshift-marketplace/redhat-marketplace-5658z" Mar 17 10:20:09 crc kubenswrapper[4813]: I0317 10:20:09.407906 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxswz\" (UniqueName: \"kubernetes.io/projected/08aaed36-4ec7-4a1e-9a9b-fc78eb29726b-kube-api-access-gxswz\") pod \"redhat-marketplace-5658z\" (UID: \"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b\") " pod="openshift-marketplace/redhat-marketplace-5658z" Mar 17 10:20:09 crc kubenswrapper[4813]: I0317 10:20:09.437145 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5658z" Mar 17 10:20:09 crc kubenswrapper[4813]: I0317 10:20:09.978149 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5658z"] Mar 17 10:20:10 crc kubenswrapper[4813]: I0317 10:20:10.809173 4813 generic.go:334] "Generic (PLEG): container finished" podID="08aaed36-4ec7-4a1e-9a9b-fc78eb29726b" containerID="b104b3652491a671fa2bafdbfbef4976c9101e7d25def77b6bbea953b747d8d0" exitCode=0 Mar 17 10:20:10 crc kubenswrapper[4813]: I0317 10:20:10.809245 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5658z" event={"ID":"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b","Type":"ContainerDied","Data":"b104b3652491a671fa2bafdbfbef4976c9101e7d25def77b6bbea953b747d8d0"} Mar 17 10:20:10 crc kubenswrapper[4813]: I0317 10:20:10.809457 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5658z" event={"ID":"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b","Type":"ContainerStarted","Data":"f60578ad24f1b880d54586f23adb05c143217dfea01644da75ba2b8cbbce61f1"} Mar 17 10:20:11 crc kubenswrapper[4813]: I0317 10:20:11.510863 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wgcbq"] Mar 17 10:20:11 crc kubenswrapper[4813]: I0317 10:20:11.515672 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wgcbq" Mar 17 10:20:11 crc kubenswrapper[4813]: I0317 10:20:11.549823 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wgcbq"] Mar 17 10:20:11 crc kubenswrapper[4813]: I0317 10:20:11.633738 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23fb3536-e50f-44dd-8aaa-28971e18c7f7-utilities\") pod \"certified-operators-wgcbq\" (UID: \"23fb3536-e50f-44dd-8aaa-28971e18c7f7\") " pod="openshift-marketplace/certified-operators-wgcbq" Mar 17 10:20:11 crc kubenswrapper[4813]: I0317 10:20:11.633894 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23fb3536-e50f-44dd-8aaa-28971e18c7f7-catalog-content\") pod \"certified-operators-wgcbq\" (UID: \"23fb3536-e50f-44dd-8aaa-28971e18c7f7\") " pod="openshift-marketplace/certified-operators-wgcbq" Mar 17 10:20:11 crc kubenswrapper[4813]: I0317 10:20:11.634194 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5sllx\" (UniqueName: \"kubernetes.io/projected/23fb3536-e50f-44dd-8aaa-28971e18c7f7-kube-api-access-5sllx\") pod \"certified-operators-wgcbq\" (UID: \"23fb3536-e50f-44dd-8aaa-28971e18c7f7\") " pod="openshift-marketplace/certified-operators-wgcbq" Mar 17 10:20:11 crc kubenswrapper[4813]: I0317 10:20:11.735637 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23fb3536-e50f-44dd-8aaa-28971e18c7f7-catalog-content\") pod \"certified-operators-wgcbq\" (UID: \"23fb3536-e50f-44dd-8aaa-28971e18c7f7\") " pod="openshift-marketplace/certified-operators-wgcbq" Mar 17 10:20:11 crc kubenswrapper[4813]: I0317 10:20:11.735777 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5sllx\" (UniqueName: \"kubernetes.io/projected/23fb3536-e50f-44dd-8aaa-28971e18c7f7-kube-api-access-5sllx\") pod \"certified-operators-wgcbq\" (UID: \"23fb3536-e50f-44dd-8aaa-28971e18c7f7\") " pod="openshift-marketplace/certified-operators-wgcbq" Mar 17 10:20:11 crc kubenswrapper[4813]: I0317 10:20:11.735896 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23fb3536-e50f-44dd-8aaa-28971e18c7f7-utilities\") pod \"certified-operators-wgcbq\" (UID: \"23fb3536-e50f-44dd-8aaa-28971e18c7f7\") " pod="openshift-marketplace/certified-operators-wgcbq" Mar 17 10:20:11 crc kubenswrapper[4813]: I0317 10:20:11.736183 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23fb3536-e50f-44dd-8aaa-28971e18c7f7-catalog-content\") pod \"certified-operators-wgcbq\" (UID: \"23fb3536-e50f-44dd-8aaa-28971e18c7f7\") " pod="openshift-marketplace/certified-operators-wgcbq" Mar 17 10:20:11 crc kubenswrapper[4813]: I0317 10:20:11.736393 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23fb3536-e50f-44dd-8aaa-28971e18c7f7-utilities\") pod \"certified-operators-wgcbq\" (UID: \"23fb3536-e50f-44dd-8aaa-28971e18c7f7\") " pod="openshift-marketplace/certified-operators-wgcbq" Mar 17 10:20:11 crc kubenswrapper[4813]: I0317 10:20:11.770689 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5sllx\" (UniqueName: \"kubernetes.io/projected/23fb3536-e50f-44dd-8aaa-28971e18c7f7-kube-api-access-5sllx\") pod \"certified-operators-wgcbq\" (UID: \"23fb3536-e50f-44dd-8aaa-28971e18c7f7\") " pod="openshift-marketplace/certified-operators-wgcbq" Mar 17 10:20:11 crc kubenswrapper[4813]: I0317 10:20:11.828153 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5658z" event={"ID":"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b","Type":"ContainerStarted","Data":"4659559d2f18478e1fdd3b18081dd1e1cd47dd8b557f2804c6886f971764d228"} Mar 17 10:20:11 crc kubenswrapper[4813]: I0317 10:20:11.879192 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wgcbq" Mar 17 10:20:12 crc kubenswrapper[4813]: I0317 10:20:12.411480 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wgcbq"] Mar 17 10:20:12 crc kubenswrapper[4813]: I0317 10:20:12.846397 4813 generic.go:334] "Generic (PLEG): container finished" podID="08aaed36-4ec7-4a1e-9a9b-fc78eb29726b" containerID="4659559d2f18478e1fdd3b18081dd1e1cd47dd8b557f2804c6886f971764d228" exitCode=0 Mar 17 10:20:12 crc kubenswrapper[4813]: I0317 10:20:12.846512 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5658z" event={"ID":"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b","Type":"ContainerDied","Data":"4659559d2f18478e1fdd3b18081dd1e1cd47dd8b557f2804c6886f971764d228"} Mar 17 10:20:12 crc kubenswrapper[4813]: I0317 10:20:12.850006 4813 generic.go:334] "Generic (PLEG): container finished" podID="23fb3536-e50f-44dd-8aaa-28971e18c7f7" containerID="2f4fc02da7540d9c9443883ab937e70cc21eb19bf0191eb1f72a89fc8952d21d" exitCode=0 Mar 17 10:20:12 crc kubenswrapper[4813]: I0317 10:20:12.850040 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgcbq" event={"ID":"23fb3536-e50f-44dd-8aaa-28971e18c7f7","Type":"ContainerDied","Data":"2f4fc02da7540d9c9443883ab937e70cc21eb19bf0191eb1f72a89fc8952d21d"} Mar 17 10:20:12 crc kubenswrapper[4813]: I0317 10:20:12.850058 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgcbq" event={"ID":"23fb3536-e50f-44dd-8aaa-28971e18c7f7","Type":"ContainerStarted","Data":"244fa5115c8407c2d9c777e9b7af5869336b998dbf50b6886618c55e235f8d47"} Mar 17 10:20:13 crc kubenswrapper[4813]: I0317 10:20:13.296644 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-858654f9db-knkd2_08c8f504-aa0d-47b9-ab86-a4b440c6f446/cert-manager-controller/0.log" Mar 17 10:20:13 crc kubenswrapper[4813]: I0317 10:20:13.529984 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-cf98fcc89-slrw5_7ff1e0fd-1a34-4f70-97c0-9fbb47f0b52b/cert-manager-cainjector/0.log" Mar 17 10:20:13 crc kubenswrapper[4813]: I0317 10:20:13.565255 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-687f57d79b-hdhpn_d949287d-f3e6-4ac8-8c16-a0d0652ae302/cert-manager-webhook/0.log" Mar 17 10:20:13 crc kubenswrapper[4813]: I0317 10:20:13.870107 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5658z" event={"ID":"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b","Type":"ContainerStarted","Data":"f4b632de16c6b909e5c8f80e145b73875eb3db89e8ce54cc1e9d16cc16e21f6e"} Mar 17 10:20:13 crc kubenswrapper[4813]: I0317 10:20:13.893480 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5658z" podStartSLOduration=2.447038602 podStartE2EDuration="4.893462626s" podCreationTimestamp="2026-03-17 10:20:09 +0000 UTC" firstStartedPulling="2026-03-17 10:20:10.811260879 +0000 UTC m=+4232.912064378" lastFinishedPulling="2026-03-17 10:20:13.257684883 +0000 UTC m=+4235.358488402" observedRunningTime="2026-03-17 10:20:13.886046965 +0000 UTC m=+4235.986850464" watchObservedRunningTime="2026-03-17 10:20:13.893462626 +0000 UTC m=+4235.994266125" Mar 17 10:20:14 crc kubenswrapper[4813]: I0317 10:20:14.881967 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgcbq" event={"ID":"23fb3536-e50f-44dd-8aaa-28971e18c7f7","Type":"ContainerStarted","Data":"1ffc038e19bc115ee19d36323ca7948639946bc1e1b6a5349f4e1cbe658cee65"} Mar 17 10:20:15 crc kubenswrapper[4813]: I0317 10:20:15.893551 4813 generic.go:334] "Generic (PLEG): container finished" podID="23fb3536-e50f-44dd-8aaa-28971e18c7f7" containerID="1ffc038e19bc115ee19d36323ca7948639946bc1e1b6a5349f4e1cbe658cee65" exitCode=0 Mar 17 10:20:15 crc kubenswrapper[4813]: I0317 10:20:15.893713 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgcbq" event={"ID":"23fb3536-e50f-44dd-8aaa-28971e18c7f7","Type":"ContainerDied","Data":"1ffc038e19bc115ee19d36323ca7948639946bc1e1b6a5349f4e1cbe658cee65"} Mar 17 10:20:16 crc kubenswrapper[4813]: I0317 10:20:16.904986 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgcbq" event={"ID":"23fb3536-e50f-44dd-8aaa-28971e18c7f7","Type":"ContainerStarted","Data":"b026f5f4ddc76fdb9da89fae87f0dd04000104974cb9322733ca3eefabe20dc2"} Mar 17 10:20:16 crc kubenswrapper[4813]: I0317 10:20:16.940578 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wgcbq" podStartSLOduration=2.392439652 podStartE2EDuration="5.940550181s" podCreationTimestamp="2026-03-17 10:20:11 +0000 UTC" firstStartedPulling="2026-03-17 10:20:12.852425856 +0000 UTC m=+4234.953229355" lastFinishedPulling="2026-03-17 10:20:16.400536385 +0000 UTC m=+4238.501339884" observedRunningTime="2026-03-17 10:20:16.929466206 +0000 UTC m=+4239.030269705" watchObservedRunningTime="2026-03-17 10:20:16.940550181 +0000 UTC m=+4239.041353710" Mar 17 10:20:19 crc kubenswrapper[4813]: I0317 10:20:19.438219 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5658z" Mar 17 10:20:19 crc kubenswrapper[4813]: I0317 10:20:19.438773 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5658z" Mar 17 10:20:19 crc kubenswrapper[4813]: I0317 10:20:19.510545 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5658z" Mar 17 10:20:19 crc kubenswrapper[4813]: I0317 10:20:19.979327 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5658z" Mar 17 10:20:21 crc kubenswrapper[4813]: I0317 10:20:21.089213 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5658z"] Mar 17 10:20:21 crc kubenswrapper[4813]: I0317 10:20:21.880169 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wgcbq" Mar 17 10:20:21 crc kubenswrapper[4813]: I0317 10:20:21.880220 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wgcbq" Mar 17 10:20:21 crc kubenswrapper[4813]: I0317 10:20:21.946929 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-5658z" podUID="08aaed36-4ec7-4a1e-9a9b-fc78eb29726b" containerName="registry-server" containerID="cri-o://f4b632de16c6b909e5c8f80e145b73875eb3db89e8ce54cc1e9d16cc16e21f6e" gracePeriod=2 Mar 17 10:20:21 crc kubenswrapper[4813]: I0317 10:20:21.948236 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wgcbq" Mar 17 10:20:22 crc kubenswrapper[4813]: I0317 10:20:22.021236 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wgcbq" Mar 17 10:20:22 crc kubenswrapper[4813]: I0317 10:20:22.436004 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5658z" Mar 17 10:20:22 crc kubenswrapper[4813]: I0317 10:20:22.543401 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08aaed36-4ec7-4a1e-9a9b-fc78eb29726b-catalog-content\") pod \"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b\" (UID: \"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b\") " Mar 17 10:20:22 crc kubenswrapper[4813]: I0317 10:20:22.543483 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxswz\" (UniqueName: \"kubernetes.io/projected/08aaed36-4ec7-4a1e-9a9b-fc78eb29726b-kube-api-access-gxswz\") pod \"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b\" (UID: \"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b\") " Mar 17 10:20:22 crc kubenswrapper[4813]: I0317 10:20:22.543651 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08aaed36-4ec7-4a1e-9a9b-fc78eb29726b-utilities\") pod \"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b\" (UID: \"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b\") " Mar 17 10:20:22 crc kubenswrapper[4813]: I0317 10:20:22.544826 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08aaed36-4ec7-4a1e-9a9b-fc78eb29726b-utilities" (OuterVolumeSpecName: "utilities") pod "08aaed36-4ec7-4a1e-9a9b-fc78eb29726b" (UID: "08aaed36-4ec7-4a1e-9a9b-fc78eb29726b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:20:22 crc kubenswrapper[4813]: I0317 10:20:22.554778 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08aaed36-4ec7-4a1e-9a9b-fc78eb29726b-kube-api-access-gxswz" (OuterVolumeSpecName: "kube-api-access-gxswz") pod "08aaed36-4ec7-4a1e-9a9b-fc78eb29726b" (UID: "08aaed36-4ec7-4a1e-9a9b-fc78eb29726b"). InnerVolumeSpecName "kube-api-access-gxswz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:20:22 crc kubenswrapper[4813]: I0317 10:20:22.589773 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/08aaed36-4ec7-4a1e-9a9b-fc78eb29726b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "08aaed36-4ec7-4a1e-9a9b-fc78eb29726b" (UID: "08aaed36-4ec7-4a1e-9a9b-fc78eb29726b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:20:22 crc kubenswrapper[4813]: I0317 10:20:22.646501 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/08aaed36-4ec7-4a1e-9a9b-fc78eb29726b-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 10:20:22 crc kubenswrapper[4813]: I0317 10:20:22.646546 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/08aaed36-4ec7-4a1e-9a9b-fc78eb29726b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 10:20:22 crc kubenswrapper[4813]: I0317 10:20:22.646570 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxswz\" (UniqueName: \"kubernetes.io/projected/08aaed36-4ec7-4a1e-9a9b-fc78eb29726b-kube-api-access-gxswz\") on node \"crc\" DevicePath \"\"" Mar 17 10:20:22 crc kubenswrapper[4813]: I0317 10:20:22.962085 4813 generic.go:334] "Generic (PLEG): container finished" podID="08aaed36-4ec7-4a1e-9a9b-fc78eb29726b" containerID="f4b632de16c6b909e5c8f80e145b73875eb3db89e8ce54cc1e9d16cc16e21f6e" exitCode=0 Mar 17 10:20:22 crc kubenswrapper[4813]: I0317 10:20:22.962155 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5658z" Mar 17 10:20:22 crc kubenswrapper[4813]: I0317 10:20:22.962168 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5658z" event={"ID":"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b","Type":"ContainerDied","Data":"f4b632de16c6b909e5c8f80e145b73875eb3db89e8ce54cc1e9d16cc16e21f6e"} Mar 17 10:20:22 crc kubenswrapper[4813]: I0317 10:20:22.962464 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5658z" event={"ID":"08aaed36-4ec7-4a1e-9a9b-fc78eb29726b","Type":"ContainerDied","Data":"f60578ad24f1b880d54586f23adb05c143217dfea01644da75ba2b8cbbce61f1"} Mar 17 10:20:22 crc kubenswrapper[4813]: I0317 10:20:22.962516 4813 scope.go:117] "RemoveContainer" containerID="f4b632de16c6b909e5c8f80e145b73875eb3db89e8ce54cc1e9d16cc16e21f6e" Mar 17 10:20:22 crc kubenswrapper[4813]: I0317 10:20:22.994963 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-5658z"] Mar 17 10:20:22 crc kubenswrapper[4813]: I0317 10:20:22.999452 4813 scope.go:117] "RemoveContainer" containerID="4659559d2f18478e1fdd3b18081dd1e1cd47dd8b557f2804c6886f971764d228" Mar 17 10:20:23 crc kubenswrapper[4813]: I0317 10:20:23.004707 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-5658z"] Mar 17 10:20:23 crc kubenswrapper[4813]: I0317 10:20:23.024621 4813 scope.go:117] "RemoveContainer" containerID="b104b3652491a671fa2bafdbfbef4976c9101e7d25def77b6bbea953b747d8d0" Mar 17 10:20:23 crc kubenswrapper[4813]: I0317 10:20:23.086396 4813 scope.go:117] "RemoveContainer" containerID="f4b632de16c6b909e5c8f80e145b73875eb3db89e8ce54cc1e9d16cc16e21f6e" Mar 17 10:20:23 crc kubenswrapper[4813]: E0317 10:20:23.087122 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4b632de16c6b909e5c8f80e145b73875eb3db89e8ce54cc1e9d16cc16e21f6e\": container with ID starting with f4b632de16c6b909e5c8f80e145b73875eb3db89e8ce54cc1e9d16cc16e21f6e not found: ID does not exist" containerID="f4b632de16c6b909e5c8f80e145b73875eb3db89e8ce54cc1e9d16cc16e21f6e" Mar 17 10:20:23 crc kubenswrapper[4813]: I0317 10:20:23.087175 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4b632de16c6b909e5c8f80e145b73875eb3db89e8ce54cc1e9d16cc16e21f6e"} err="failed to get container status \"f4b632de16c6b909e5c8f80e145b73875eb3db89e8ce54cc1e9d16cc16e21f6e\": rpc error: code = NotFound desc = could not find container \"f4b632de16c6b909e5c8f80e145b73875eb3db89e8ce54cc1e9d16cc16e21f6e\": container with ID starting with f4b632de16c6b909e5c8f80e145b73875eb3db89e8ce54cc1e9d16cc16e21f6e not found: ID does not exist" Mar 17 10:20:23 crc kubenswrapper[4813]: I0317 10:20:23.087207 4813 scope.go:117] "RemoveContainer" containerID="4659559d2f18478e1fdd3b18081dd1e1cd47dd8b557f2804c6886f971764d228" Mar 17 10:20:23 crc kubenswrapper[4813]: E0317 10:20:23.087468 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4659559d2f18478e1fdd3b18081dd1e1cd47dd8b557f2804c6886f971764d228\": container with ID starting with 4659559d2f18478e1fdd3b18081dd1e1cd47dd8b557f2804c6886f971764d228 not found: ID does not exist" containerID="4659559d2f18478e1fdd3b18081dd1e1cd47dd8b557f2804c6886f971764d228" Mar 17 10:20:23 crc kubenswrapper[4813]: I0317 10:20:23.087495 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4659559d2f18478e1fdd3b18081dd1e1cd47dd8b557f2804c6886f971764d228"} err="failed to get container status \"4659559d2f18478e1fdd3b18081dd1e1cd47dd8b557f2804c6886f971764d228\": rpc error: code = NotFound desc = could not find container \"4659559d2f18478e1fdd3b18081dd1e1cd47dd8b557f2804c6886f971764d228\": container with ID starting with 4659559d2f18478e1fdd3b18081dd1e1cd47dd8b557f2804c6886f971764d228 not found: ID does not exist" Mar 17 10:20:23 crc kubenswrapper[4813]: I0317 10:20:23.087512 4813 scope.go:117] "RemoveContainer" containerID="b104b3652491a671fa2bafdbfbef4976c9101e7d25def77b6bbea953b747d8d0" Mar 17 10:20:23 crc kubenswrapper[4813]: E0317 10:20:23.087738 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b104b3652491a671fa2bafdbfbef4976c9101e7d25def77b6bbea953b747d8d0\": container with ID starting with b104b3652491a671fa2bafdbfbef4976c9101e7d25def77b6bbea953b747d8d0 not found: ID does not exist" containerID="b104b3652491a671fa2bafdbfbef4976c9101e7d25def77b6bbea953b747d8d0" Mar 17 10:20:23 crc kubenswrapper[4813]: I0317 10:20:23.087764 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b104b3652491a671fa2bafdbfbef4976c9101e7d25def77b6bbea953b747d8d0"} err="failed to get container status \"b104b3652491a671fa2bafdbfbef4976c9101e7d25def77b6bbea953b747d8d0\": rpc error: code = NotFound desc = could not find container \"b104b3652491a671fa2bafdbfbef4976c9101e7d25def77b6bbea953b747d8d0\": container with ID starting with b104b3652491a671fa2bafdbfbef4976c9101e7d25def77b6bbea953b747d8d0 not found: ID does not exist" Mar 17 10:20:24 crc kubenswrapper[4813]: I0317 10:20:24.284888 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wgcbq"] Mar 17 10:20:24 crc kubenswrapper[4813]: I0317 10:20:24.286171 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wgcbq" podUID="23fb3536-e50f-44dd-8aaa-28971e18c7f7" containerName="registry-server" containerID="cri-o://b026f5f4ddc76fdb9da89fae87f0dd04000104974cb9322733ca3eefabe20dc2" gracePeriod=2 Mar 17 10:20:24 crc kubenswrapper[4813]: I0317 10:20:24.740887 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08aaed36-4ec7-4a1e-9a9b-fc78eb29726b" path="/var/lib/kubelet/pods/08aaed36-4ec7-4a1e-9a9b-fc78eb29726b/volumes" Mar 17 10:20:24 crc kubenswrapper[4813]: I0317 10:20:24.778750 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wgcbq" Mar 17 10:20:24 crc kubenswrapper[4813]: I0317 10:20:24.891628 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23fb3536-e50f-44dd-8aaa-28971e18c7f7-utilities\") pod \"23fb3536-e50f-44dd-8aaa-28971e18c7f7\" (UID: \"23fb3536-e50f-44dd-8aaa-28971e18c7f7\") " Mar 17 10:20:24 crc kubenswrapper[4813]: I0317 10:20:24.891770 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23fb3536-e50f-44dd-8aaa-28971e18c7f7-catalog-content\") pod \"23fb3536-e50f-44dd-8aaa-28971e18c7f7\" (UID: \"23fb3536-e50f-44dd-8aaa-28971e18c7f7\") " Mar 17 10:20:24 crc kubenswrapper[4813]: I0317 10:20:24.891925 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5sllx\" (UniqueName: \"kubernetes.io/projected/23fb3536-e50f-44dd-8aaa-28971e18c7f7-kube-api-access-5sllx\") pod \"23fb3536-e50f-44dd-8aaa-28971e18c7f7\" (UID: \"23fb3536-e50f-44dd-8aaa-28971e18c7f7\") " Mar 17 10:20:24 crc kubenswrapper[4813]: I0317 10:20:24.892921 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23fb3536-e50f-44dd-8aaa-28971e18c7f7-utilities" (OuterVolumeSpecName: "utilities") pod "23fb3536-e50f-44dd-8aaa-28971e18c7f7" (UID: "23fb3536-e50f-44dd-8aaa-28971e18c7f7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:20:24 crc kubenswrapper[4813]: I0317 10:20:24.897469 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23fb3536-e50f-44dd-8aaa-28971e18c7f7-kube-api-access-5sllx" (OuterVolumeSpecName: "kube-api-access-5sllx") pod "23fb3536-e50f-44dd-8aaa-28971e18c7f7" (UID: "23fb3536-e50f-44dd-8aaa-28971e18c7f7"). InnerVolumeSpecName "kube-api-access-5sllx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:20:24 crc kubenswrapper[4813]: I0317 10:20:24.960699 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/23fb3536-e50f-44dd-8aaa-28971e18c7f7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "23fb3536-e50f-44dd-8aaa-28971e18c7f7" (UID: "23fb3536-e50f-44dd-8aaa-28971e18c7f7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:20:24 crc kubenswrapper[4813]: I0317 10:20:24.988382 4813 generic.go:334] "Generic (PLEG): container finished" podID="23fb3536-e50f-44dd-8aaa-28971e18c7f7" containerID="b026f5f4ddc76fdb9da89fae87f0dd04000104974cb9322733ca3eefabe20dc2" exitCode=0 Mar 17 10:20:24 crc kubenswrapper[4813]: I0317 10:20:24.988438 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgcbq" event={"ID":"23fb3536-e50f-44dd-8aaa-28971e18c7f7","Type":"ContainerDied","Data":"b026f5f4ddc76fdb9da89fae87f0dd04000104974cb9322733ca3eefabe20dc2"} Mar 17 10:20:24 crc kubenswrapper[4813]: I0317 10:20:24.988830 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgcbq" event={"ID":"23fb3536-e50f-44dd-8aaa-28971e18c7f7","Type":"ContainerDied","Data":"244fa5115c8407c2d9c777e9b7af5869336b998dbf50b6886618c55e235f8d47"} Mar 17 10:20:24 crc kubenswrapper[4813]: I0317 10:20:24.988950 4813 scope.go:117] "RemoveContainer" containerID="b026f5f4ddc76fdb9da89fae87f0dd04000104974cb9322733ca3eefabe20dc2" Mar 17 10:20:24 crc kubenswrapper[4813]: I0317 10:20:24.988462 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wgcbq" Mar 17 10:20:24 crc kubenswrapper[4813]: I0317 10:20:24.994360 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5sllx\" (UniqueName: \"kubernetes.io/projected/23fb3536-e50f-44dd-8aaa-28971e18c7f7-kube-api-access-5sllx\") on node \"crc\" DevicePath \"\"" Mar 17 10:20:24 crc kubenswrapper[4813]: I0317 10:20:24.994390 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/23fb3536-e50f-44dd-8aaa-28971e18c7f7-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 10:20:24 crc kubenswrapper[4813]: I0317 10:20:24.994401 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/23fb3536-e50f-44dd-8aaa-28971e18c7f7-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 10:20:25 crc kubenswrapper[4813]: I0317 10:20:25.017948 4813 scope.go:117] "RemoveContainer" containerID="1ffc038e19bc115ee19d36323ca7948639946bc1e1b6a5349f4e1cbe658cee65" Mar 17 10:20:25 crc kubenswrapper[4813]: I0317 10:20:25.024111 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wgcbq"] Mar 17 10:20:25 crc kubenswrapper[4813]: I0317 10:20:25.031514 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wgcbq"] Mar 17 10:20:25 crc kubenswrapper[4813]: I0317 10:20:25.060071 4813 scope.go:117] "RemoveContainer" containerID="2f4fc02da7540d9c9443883ab937e70cc21eb19bf0191eb1f72a89fc8952d21d" Mar 17 10:20:25 crc kubenswrapper[4813]: I0317 10:20:25.096194 4813 scope.go:117] "RemoveContainer" containerID="b026f5f4ddc76fdb9da89fae87f0dd04000104974cb9322733ca3eefabe20dc2" Mar 17 10:20:25 crc kubenswrapper[4813]: E0317 10:20:25.096837 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b026f5f4ddc76fdb9da89fae87f0dd04000104974cb9322733ca3eefabe20dc2\": container with ID starting with b026f5f4ddc76fdb9da89fae87f0dd04000104974cb9322733ca3eefabe20dc2 not found: ID does not exist" containerID="b026f5f4ddc76fdb9da89fae87f0dd04000104974cb9322733ca3eefabe20dc2" Mar 17 10:20:25 crc kubenswrapper[4813]: I0317 10:20:25.096880 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b026f5f4ddc76fdb9da89fae87f0dd04000104974cb9322733ca3eefabe20dc2"} err="failed to get container status \"b026f5f4ddc76fdb9da89fae87f0dd04000104974cb9322733ca3eefabe20dc2\": rpc error: code = NotFound desc = could not find container \"b026f5f4ddc76fdb9da89fae87f0dd04000104974cb9322733ca3eefabe20dc2\": container with ID starting with b026f5f4ddc76fdb9da89fae87f0dd04000104974cb9322733ca3eefabe20dc2 not found: ID does not exist" Mar 17 10:20:25 crc kubenswrapper[4813]: I0317 10:20:25.096934 4813 scope.go:117] "RemoveContainer" containerID="1ffc038e19bc115ee19d36323ca7948639946bc1e1b6a5349f4e1cbe658cee65" Mar 17 10:20:25 crc kubenswrapper[4813]: E0317 10:20:25.097667 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ffc038e19bc115ee19d36323ca7948639946bc1e1b6a5349f4e1cbe658cee65\": container with ID starting with 1ffc038e19bc115ee19d36323ca7948639946bc1e1b6a5349f4e1cbe658cee65 not found: ID does not exist" containerID="1ffc038e19bc115ee19d36323ca7948639946bc1e1b6a5349f4e1cbe658cee65" Mar 17 10:20:25 crc kubenswrapper[4813]: I0317 10:20:25.097786 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ffc038e19bc115ee19d36323ca7948639946bc1e1b6a5349f4e1cbe658cee65"} err="failed to get container status \"1ffc038e19bc115ee19d36323ca7948639946bc1e1b6a5349f4e1cbe658cee65\": rpc error: code = NotFound desc = could not find container \"1ffc038e19bc115ee19d36323ca7948639946bc1e1b6a5349f4e1cbe658cee65\": container with ID starting with 1ffc038e19bc115ee19d36323ca7948639946bc1e1b6a5349f4e1cbe658cee65 not found: ID does not exist" Mar 17 10:20:25 crc kubenswrapper[4813]: I0317 10:20:25.097824 4813 scope.go:117] "RemoveContainer" containerID="2f4fc02da7540d9c9443883ab937e70cc21eb19bf0191eb1f72a89fc8952d21d" Mar 17 10:20:25 crc kubenswrapper[4813]: E0317 10:20:25.098327 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f4fc02da7540d9c9443883ab937e70cc21eb19bf0191eb1f72a89fc8952d21d\": container with ID starting with 2f4fc02da7540d9c9443883ab937e70cc21eb19bf0191eb1f72a89fc8952d21d not found: ID does not exist" containerID="2f4fc02da7540d9c9443883ab937e70cc21eb19bf0191eb1f72a89fc8952d21d" Mar 17 10:20:25 crc kubenswrapper[4813]: I0317 10:20:25.098357 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f4fc02da7540d9c9443883ab937e70cc21eb19bf0191eb1f72a89fc8952d21d"} err="failed to get container status \"2f4fc02da7540d9c9443883ab937e70cc21eb19bf0191eb1f72a89fc8952d21d\": rpc error: code = NotFound desc = could not find container \"2f4fc02da7540d9c9443883ab937e70cc21eb19bf0191eb1f72a89fc8952d21d\": container with ID starting with 2f4fc02da7540d9c9443883ab937e70cc21eb19bf0191eb1f72a89fc8952d21d not found: ID does not exist" Mar 17 10:20:26 crc kubenswrapper[4813]: I0317 10:20:26.744546 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23fb3536-e50f-44dd-8aaa-28971e18c7f7" path="/var/lib/kubelet/pods/23fb3536-e50f-44dd-8aaa-28971e18c7f7/volumes" Mar 17 10:20:28 crc kubenswrapper[4813]: I0317 10:20:28.452411 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-86f58fcf4-4jmsm_89acc01c-091e-402c-a206-8fbc201f0884/nmstate-console-plugin/0.log" Mar 17 10:20:28 crc kubenswrapper[4813]: I0317 10:20:28.582613 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-zckpn_8813052b-cc40-4fe7-b4de-7aeb3832d4bb/nmstate-handler/0.log" Mar 17 10:20:28 crc kubenswrapper[4813]: I0317 10:20:28.622158 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-k2x5n_0d0f7c9a-ffef-4d0b-ae31-0cfd9e864802/kube-rbac-proxy/0.log" Mar 17 10:20:28 crc kubenswrapper[4813]: I0317 10:20:28.755408 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-9b8c8685d-k2x5n_0d0f7c9a-ffef-4d0b-ae31-0cfd9e864802/nmstate-metrics/0.log" Mar 17 10:20:28 crc kubenswrapper[4813]: I0317 10:20:28.852188 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-796d4cfff4-77p2p_acc2af0d-b40c-48d9-9b83-d0cf7755c37b/nmstate-operator/0.log" Mar 17 10:20:28 crc kubenswrapper[4813]: I0317 10:20:28.993889 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f558f5558-95n52_23256f06-964f-4f6f-bedf-6159a92a06d0/nmstate-webhook/0.log" Mar 17 10:20:51 crc kubenswrapper[4813]: I0317 10:20:51.914826 4813 scope.go:117] "RemoveContainer" containerID="9b867a204e4e466a5ab442da560f36a8efd01a81a3e34a285b48127292737f1a" Mar 17 10:20:57 crc kubenswrapper[4813]: I0317 10:20:57.990060 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-5cxrm_8503b49f-ae8d-4e0d-b3c6-30320e0e283a/kube-rbac-proxy/0.log" Mar 17 10:20:58 crc kubenswrapper[4813]: I0317 10:20:58.017761 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-7bb4cc7c98-5cxrm_8503b49f-ae8d-4e0d-b3c6-30320e0e283a/controller/0.log" Mar 17 10:20:58 crc kubenswrapper[4813]: I0317 10:20:58.205857 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-frr-files/0.log" Mar 17 10:20:58 crc kubenswrapper[4813]: I0317 10:20:58.333272 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-frr-files/0.log" Mar 17 10:20:58 crc kubenswrapper[4813]: I0317 10:20:58.377671 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-metrics/0.log" Mar 17 10:20:58 crc kubenswrapper[4813]: I0317 10:20:58.380785 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-reloader/0.log" Mar 17 10:20:58 crc kubenswrapper[4813]: I0317 10:20:58.384423 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-reloader/0.log" Mar 17 10:20:58 crc kubenswrapper[4813]: I0317 10:20:58.591150 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-metrics/0.log" Mar 17 10:20:58 crc kubenswrapper[4813]: I0317 10:20:58.593095 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-reloader/0.log" Mar 17 10:20:58 crc kubenswrapper[4813]: I0317 10:20:58.599442 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-frr-files/0.log" Mar 17 10:20:58 crc kubenswrapper[4813]: I0317 10:20:58.606182 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-metrics/0.log" Mar 17 10:20:58 crc kubenswrapper[4813]: I0317 10:20:58.787410 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-metrics/0.log" Mar 17 10:20:58 crc kubenswrapper[4813]: I0317 10:20:58.793960 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/controller/0.log" Mar 17 10:20:58 crc kubenswrapper[4813]: I0317 10:20:58.794047 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-reloader/0.log" Mar 17 10:20:58 crc kubenswrapper[4813]: I0317 10:20:58.823296 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/cp-frr-files/0.log" Mar 17 10:20:58 crc kubenswrapper[4813]: I0317 10:20:58.999505 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/kube-rbac-proxy/0.log" Mar 17 10:20:59 crc kubenswrapper[4813]: I0317 10:20:59.041892 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/kube-rbac-proxy-frr/0.log" Mar 17 10:20:59 crc kubenswrapper[4813]: I0317 10:20:59.089100 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/frr-metrics/0.log" Mar 17 10:20:59 crc kubenswrapper[4813]: I0317 10:20:59.181396 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/reloader/0.log" Mar 17 10:20:59 crc kubenswrapper[4813]: I0317 10:20:59.264065 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-bcc4b6f68-wjgfj_30c2f61b-c33a-49c4-bde0-ffe8593f29b8/frr-k8s-webhook-server/0.log" Mar 17 10:20:59 crc kubenswrapper[4813]: I0317 10:20:59.427030 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-56666fbc9b-dj55n_a63b7bd8-1a98-40d9-9f56-c7662d229b3d/manager/0.log" Mar 17 10:20:59 crc kubenswrapper[4813]: I0317 10:20:59.561385 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-55494b4d7-h6xch_f83986c8-b147-48b9-84ec-f09d01ddd8ce/webhook-server/0.log" Mar 17 10:20:59 crc kubenswrapper[4813]: I0317 10:20:59.674876 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hzk8t_ece7eee8-f37e-43f4-a762-986d7d1ae876/kube-rbac-proxy/0.log" Mar 17 10:21:00 crc kubenswrapper[4813]: I0317 10:21:00.272376 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-hzk8t_ece7eee8-f37e-43f4-a762-986d7d1ae876/speaker/0.log" Mar 17 10:21:00 crc kubenswrapper[4813]: I0317 10:21:00.692685 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-hj2sl_fbcdac92-95fa-4d87-bc00-fccd03f4c660/frr/0.log" Mar 17 10:21:16 crc kubenswrapper[4813]: I0317 10:21:16.109834 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf_bb556790-2a96-412b-9847-3cd62873bfed/util/0.log" Mar 17 10:21:16 crc kubenswrapper[4813]: I0317 10:21:16.259913 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf_bb556790-2a96-412b-9847-3cd62873bfed/pull/0.log" Mar 17 10:21:16 crc kubenswrapper[4813]: I0317 10:21:16.297193 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf_bb556790-2a96-412b-9847-3cd62873bfed/util/0.log" Mar 17 10:21:16 crc kubenswrapper[4813]: I0317 10:21:16.306955 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf_bb556790-2a96-412b-9847-3cd62873bfed/pull/0.log" Mar 17 10:21:16 crc kubenswrapper[4813]: I0317 10:21:16.440411 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf_bb556790-2a96-412b-9847-3cd62873bfed/util/0.log" Mar 17 10:21:16 crc kubenswrapper[4813]: I0317 10:21:16.483320 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf_bb556790-2a96-412b-9847-3cd62873bfed/extract/0.log" Mar 17 10:21:16 crc kubenswrapper[4813]: I0317 10:21:16.484568 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_1d8741a795bd73341bdd61a6e59c08511cf9466dbb5fc4045ac2dde874jpdbf_bb556790-2a96-412b-9847-3cd62873bfed/pull/0.log" Mar 17 10:21:16 crc kubenswrapper[4813]: I0317 10:21:16.631313 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98_0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2/util/0.log" Mar 17 10:21:17 crc kubenswrapper[4813]: I0317 10:21:17.286965 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98_0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2/pull/0.log" Mar 17 10:21:17 crc kubenswrapper[4813]: I0317 10:21:17.301469 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98_0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2/util/0.log" Mar 17 10:21:17 crc kubenswrapper[4813]: I0317 10:21:17.334048 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98_0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2/pull/0.log" Mar 17 10:21:17 crc kubenswrapper[4813]: I0317 10:21:17.469770 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98_0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2/extract/0.log" Mar 17 10:21:17 crc kubenswrapper[4813]: I0317 10:21:17.496617 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98_0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2/util/0.log" Mar 17 10:21:17 crc kubenswrapper[4813]: I0317 10:21:17.517518 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_2d3ddce10053cc6867b5a0ce1614b30225f3a63fab79a72148165675c1rrm98_0b47f2fe-3d7e-4550-a89f-79d3af2cc8f2/pull/0.log" Mar 17 10:21:17 crc kubenswrapper[4813]: I0317 10:21:17.641168 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q2m6c_8840c430-4827-4da1-baec-38cff419ed36/extract-utilities/0.log" Mar 17 10:21:17 crc kubenswrapper[4813]: I0317 10:21:17.774322 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q2m6c_8840c430-4827-4da1-baec-38cff419ed36/extract-content/0.log" Mar 17 10:21:17 crc kubenswrapper[4813]: I0317 10:21:17.774394 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q2m6c_8840c430-4827-4da1-baec-38cff419ed36/extract-content/0.log" Mar 17 10:21:17 crc kubenswrapper[4813]: I0317 10:21:17.804724 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q2m6c_8840c430-4827-4da1-baec-38cff419ed36/extract-utilities/0.log" Mar 17 10:21:18 crc kubenswrapper[4813]: I0317 10:21:18.009694 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q2m6c_8840c430-4827-4da1-baec-38cff419ed36/extract-content/0.log" Mar 17 10:21:18 crc kubenswrapper[4813]: I0317 10:21:18.014826 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q2m6c_8840c430-4827-4da1-baec-38cff419ed36/extract-utilities/0.log" Mar 17 10:21:18 crc kubenswrapper[4813]: I0317 10:21:18.206118 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pr6tk_f8b71cf5-0aee-4d4e-a861-bb1bb52fe746/extract-utilities/0.log" Mar 17 10:21:18 crc kubenswrapper[4813]: I0317 10:21:18.340068 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-q2m6c_8840c430-4827-4da1-baec-38cff419ed36/registry-server/0.log" Mar 17 10:21:18 crc kubenswrapper[4813]: I0317 10:21:18.438147 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pr6tk_f8b71cf5-0aee-4d4e-a861-bb1bb52fe746/extract-utilities/0.log" Mar 17 10:21:18 crc kubenswrapper[4813]: I0317 10:21:18.441073 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pr6tk_f8b71cf5-0aee-4d4e-a861-bb1bb52fe746/extract-content/0.log" Mar 17 10:21:18 crc kubenswrapper[4813]: I0317 10:21:18.449894 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pr6tk_f8b71cf5-0aee-4d4e-a861-bb1bb52fe746/extract-content/0.log" Mar 17 10:21:18 crc kubenswrapper[4813]: I0317 10:21:18.622086 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pr6tk_f8b71cf5-0aee-4d4e-a861-bb1bb52fe746/extract-utilities/0.log" Mar 17 10:21:18 crc kubenswrapper[4813]: I0317 10:21:18.643028 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pr6tk_f8b71cf5-0aee-4d4e-a861-bb1bb52fe746/extract-content/0.log" Mar 17 10:21:18 crc kubenswrapper[4813]: I0317 10:21:18.884433 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-nw44l_6ca7f195-b5c8-4804-b95a-5ff53a637ac5/marketplace-operator/0.log" Mar 17 10:21:18 crc kubenswrapper[4813]: I0317 10:21:18.963658 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-54j2x_c7e12568-6cf2-4551-8aff-b330c8eda201/extract-utilities/0.log" Mar 17 10:21:19 crc kubenswrapper[4813]: I0317 10:21:19.011918 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-pr6tk_f8b71cf5-0aee-4d4e-a861-bb1bb52fe746/registry-server/0.log" Mar 17 10:21:19 crc kubenswrapper[4813]: I0317 10:21:19.119303 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-54j2x_c7e12568-6cf2-4551-8aff-b330c8eda201/extract-content/0.log" Mar 17 10:21:19 crc kubenswrapper[4813]: I0317 10:21:19.159712 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-54j2x_c7e12568-6cf2-4551-8aff-b330c8eda201/extract-content/0.log" Mar 17 10:21:19 crc kubenswrapper[4813]: I0317 10:21:19.168996 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-54j2x_c7e12568-6cf2-4551-8aff-b330c8eda201/extract-utilities/0.log" Mar 17 10:21:19 crc kubenswrapper[4813]: I0317 10:21:19.340537 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-54j2x_c7e12568-6cf2-4551-8aff-b330c8eda201/extract-utilities/0.log" Mar 17 10:21:19 crc kubenswrapper[4813]: I0317 10:21:19.370067 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-54j2x_c7e12568-6cf2-4551-8aff-b330c8eda201/extract-content/0.log" Mar 17 10:21:19 crc kubenswrapper[4813]: I0317 10:21:19.374995 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vf8rq_4fc7a47b-182b-40dc-bf52-752b7434f448/extract-utilities/0.log" Mar 17 10:21:19 crc kubenswrapper[4813]: I0317 10:21:19.430111 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-54j2x_c7e12568-6cf2-4551-8aff-b330c8eda201/registry-server/0.log" Mar 17 10:21:19 crc kubenswrapper[4813]: I0317 10:21:19.549650 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vf8rq_4fc7a47b-182b-40dc-bf52-752b7434f448/extract-content/0.log" Mar 17 10:21:19 crc kubenswrapper[4813]: I0317 10:21:19.557049 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vf8rq_4fc7a47b-182b-40dc-bf52-752b7434f448/extract-utilities/0.log" Mar 17 10:21:19 crc kubenswrapper[4813]: I0317 10:21:19.590771 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vf8rq_4fc7a47b-182b-40dc-bf52-752b7434f448/extract-content/0.log" Mar 17 10:21:19 crc kubenswrapper[4813]: I0317 10:21:19.739080 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vf8rq_4fc7a47b-182b-40dc-bf52-752b7434f448/extract-content/0.log" Mar 17 10:21:19 crc kubenswrapper[4813]: I0317 10:21:19.777375 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vf8rq_4fc7a47b-182b-40dc-bf52-752b7434f448/extract-utilities/0.log" Mar 17 10:21:20 crc kubenswrapper[4813]: I0317 10:21:20.241768 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vf8rq_4fc7a47b-182b-40dc-bf52-752b7434f448/registry-server/0.log" Mar 17 10:21:44 crc kubenswrapper[4813]: I0317 10:21:44.113506 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 10:21:44 crc kubenswrapper[4813]: I0317 10:21:44.114121 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 10:22:00 crc kubenswrapper[4813]: I0317 10:22:00.171204 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562382-c7cjk"] Mar 17 10:22:00 crc kubenswrapper[4813]: E0317 10:22:00.174671 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23fb3536-e50f-44dd-8aaa-28971e18c7f7" containerName="extract-utilities" Mar 17 10:22:00 crc kubenswrapper[4813]: I0317 10:22:00.174867 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="23fb3536-e50f-44dd-8aaa-28971e18c7f7" containerName="extract-utilities" Mar 17 10:22:00 crc kubenswrapper[4813]: E0317 10:22:00.175024 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08aaed36-4ec7-4a1e-9a9b-fc78eb29726b" containerName="extract-content" Mar 17 10:22:00 crc kubenswrapper[4813]: I0317 10:22:00.175160 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="08aaed36-4ec7-4a1e-9a9b-fc78eb29726b" containerName="extract-content" Mar 17 10:22:00 crc kubenswrapper[4813]: E0317 10:22:00.175295 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23fb3536-e50f-44dd-8aaa-28971e18c7f7" containerName="extract-content" Mar 17 10:22:00 crc kubenswrapper[4813]: I0317 10:22:00.175434 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="23fb3536-e50f-44dd-8aaa-28971e18c7f7" containerName="extract-content" Mar 17 10:22:00 crc kubenswrapper[4813]: E0317 10:22:00.175650 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08aaed36-4ec7-4a1e-9a9b-fc78eb29726b" containerName="registry-server" Mar 17 10:22:00 crc kubenswrapper[4813]: I0317 10:22:00.175802 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="08aaed36-4ec7-4a1e-9a9b-fc78eb29726b" containerName="registry-server" Mar 17 10:22:00 crc kubenswrapper[4813]: E0317 10:22:00.175951 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08aaed36-4ec7-4a1e-9a9b-fc78eb29726b" containerName="extract-utilities" Mar 17 10:22:00 crc kubenswrapper[4813]: I0317 10:22:00.183024 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="08aaed36-4ec7-4a1e-9a9b-fc78eb29726b" containerName="extract-utilities" Mar 17 10:22:00 crc kubenswrapper[4813]: E0317 10:22:00.183332 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23fb3536-e50f-44dd-8aaa-28971e18c7f7" containerName="registry-server" Mar 17 10:22:00 crc kubenswrapper[4813]: I0317 10:22:00.183465 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="23fb3536-e50f-44dd-8aaa-28971e18c7f7" containerName="registry-server" Mar 17 10:22:00 crc kubenswrapper[4813]: I0317 10:22:00.184071 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="08aaed36-4ec7-4a1e-9a9b-fc78eb29726b" containerName="registry-server" Mar 17 10:22:00 crc kubenswrapper[4813]: I0317 10:22:00.184218 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="23fb3536-e50f-44dd-8aaa-28971e18c7f7" containerName="registry-server" Mar 17 10:22:00 crc kubenswrapper[4813]: I0317 10:22:00.185068 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562382-c7cjk"] Mar 17 10:22:00 crc kubenswrapper[4813]: I0317 10:22:00.185269 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562382-c7cjk" Mar 17 10:22:00 crc kubenswrapper[4813]: I0317 10:22:00.193354 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 10:22:00 crc kubenswrapper[4813]: I0317 10:22:00.194357 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 10:22:00 crc kubenswrapper[4813]: I0317 10:22:00.194770 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 10:22:00 crc kubenswrapper[4813]: I0317 10:22:00.224019 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qqvn\" (UniqueName: \"kubernetes.io/projected/ca931d57-2657-4c46-bb87-4e3fa445f2ed-kube-api-access-5qqvn\") pod \"auto-csr-approver-29562382-c7cjk\" (UID: \"ca931d57-2657-4c46-bb87-4e3fa445f2ed\") " pod="openshift-infra/auto-csr-approver-29562382-c7cjk" Mar 17 10:22:00 crc kubenswrapper[4813]: I0317 10:22:00.596781 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qqvn\" (UniqueName: \"kubernetes.io/projected/ca931d57-2657-4c46-bb87-4e3fa445f2ed-kube-api-access-5qqvn\") pod \"auto-csr-approver-29562382-c7cjk\" (UID: \"ca931d57-2657-4c46-bb87-4e3fa445f2ed\") " pod="openshift-infra/auto-csr-approver-29562382-c7cjk" Mar 17 10:22:00 crc kubenswrapper[4813]: I0317 10:22:00.618480 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qqvn\" (UniqueName: \"kubernetes.io/projected/ca931d57-2657-4c46-bb87-4e3fa445f2ed-kube-api-access-5qqvn\") pod \"auto-csr-approver-29562382-c7cjk\" (UID: \"ca931d57-2657-4c46-bb87-4e3fa445f2ed\") " pod="openshift-infra/auto-csr-approver-29562382-c7cjk" Mar 17 10:22:00 crc kubenswrapper[4813]: I0317 10:22:00.811688 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562382-c7cjk" Mar 17 10:22:01 crc kubenswrapper[4813]: I0317 10:22:01.594076 4813 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 17 10:22:01 crc kubenswrapper[4813]: I0317 10:22:01.603758 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562382-c7cjk"] Mar 17 10:22:01 crc kubenswrapper[4813]: I0317 10:22:01.963004 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562382-c7cjk" event={"ID":"ca931d57-2657-4c46-bb87-4e3fa445f2ed","Type":"ContainerStarted","Data":"46baa511b59847a598610e5124d23ab4b10c7fa72816676333ad5699d20a7945"} Mar 17 10:22:03 crc kubenswrapper[4813]: I0317 10:22:03.990860 4813 generic.go:334] "Generic (PLEG): container finished" podID="ca931d57-2657-4c46-bb87-4e3fa445f2ed" containerID="493d9c77cf58e2733c97abf5e931d840f511a0d880fde1a80db74fc0e967361b" exitCode=0 Mar 17 10:22:03 crc kubenswrapper[4813]: I0317 10:22:03.990996 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562382-c7cjk" event={"ID":"ca931d57-2657-4c46-bb87-4e3fa445f2ed","Type":"ContainerDied","Data":"493d9c77cf58e2733c97abf5e931d840f511a0d880fde1a80db74fc0e967361b"} Mar 17 10:22:05 crc kubenswrapper[4813]: I0317 10:22:05.396436 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562382-c7cjk" Mar 17 10:22:05 crc kubenswrapper[4813]: I0317 10:22:05.489950 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qqvn\" (UniqueName: \"kubernetes.io/projected/ca931d57-2657-4c46-bb87-4e3fa445f2ed-kube-api-access-5qqvn\") pod \"ca931d57-2657-4c46-bb87-4e3fa445f2ed\" (UID: \"ca931d57-2657-4c46-bb87-4e3fa445f2ed\") " Mar 17 10:22:05 crc kubenswrapper[4813]: I0317 10:22:05.498675 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca931d57-2657-4c46-bb87-4e3fa445f2ed-kube-api-access-5qqvn" (OuterVolumeSpecName: "kube-api-access-5qqvn") pod "ca931d57-2657-4c46-bb87-4e3fa445f2ed" (UID: "ca931d57-2657-4c46-bb87-4e3fa445f2ed"). InnerVolumeSpecName "kube-api-access-5qqvn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:22:05 crc kubenswrapper[4813]: I0317 10:22:05.592387 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qqvn\" (UniqueName: \"kubernetes.io/projected/ca931d57-2657-4c46-bb87-4e3fa445f2ed-kube-api-access-5qqvn\") on node \"crc\" DevicePath \"\"" Mar 17 10:22:06 crc kubenswrapper[4813]: I0317 10:22:06.012393 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562382-c7cjk" event={"ID":"ca931d57-2657-4c46-bb87-4e3fa445f2ed","Type":"ContainerDied","Data":"46baa511b59847a598610e5124d23ab4b10c7fa72816676333ad5699d20a7945"} Mar 17 10:22:06 crc kubenswrapper[4813]: I0317 10:22:06.012716 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="46baa511b59847a598610e5124d23ab4b10c7fa72816676333ad5699d20a7945" Mar 17 10:22:06 crc kubenswrapper[4813]: I0317 10:22:06.012445 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562382-c7cjk" Mar 17 10:22:06 crc kubenswrapper[4813]: I0317 10:22:06.493933 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562376-c2ztw"] Mar 17 10:22:06 crc kubenswrapper[4813]: I0317 10:22:06.504970 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562376-c2ztw"] Mar 17 10:22:06 crc kubenswrapper[4813]: I0317 10:22:06.746277 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049" path="/var/lib/kubelet/pods/8a5cb7ef-6bc3-4ae2-b81d-0019d4b88049/volumes" Mar 17 10:22:14 crc kubenswrapper[4813]: I0317 10:22:14.114644 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 10:22:14 crc kubenswrapper[4813]: I0317 10:22:14.115530 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 10:22:44 crc kubenswrapper[4813]: I0317 10:22:44.114508 4813 patch_prober.go:28] interesting pod/machine-config-daemon-7xxkf container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 17 10:22:44 crc kubenswrapper[4813]: I0317 10:22:44.115425 4813 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 17 10:22:44 crc kubenswrapper[4813]: I0317 10:22:44.115508 4813 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" Mar 17 10:22:44 crc kubenswrapper[4813]: I0317 10:22:44.116891 4813 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee"} pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 17 10:22:44 crc kubenswrapper[4813]: I0317 10:22:44.117003 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerName="machine-config-daemon" containerID="cri-o://c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" gracePeriod=600 Mar 17 10:22:44 crc kubenswrapper[4813]: E0317 10:22:44.248566 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:22:44 crc kubenswrapper[4813]: I0317 10:22:44.516134 4813 generic.go:334] "Generic (PLEG): container finished" podID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" containerID="c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" exitCode=0 Mar 17 10:22:44 crc kubenswrapper[4813]: I0317 10:22:44.516204 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" event={"ID":"e7e156ce-b279-4a25-a27e-2da7d0ba6437","Type":"ContainerDied","Data":"c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee"} Mar 17 10:22:44 crc kubenswrapper[4813]: I0317 10:22:44.516252 4813 scope.go:117] "RemoveContainer" containerID="38fdc8059c00804eb8afc6b3d9dac72a950b31fe31c73eb825628373cb5b059d" Mar 17 10:22:44 crc kubenswrapper[4813]: I0317 10:22:44.517250 4813 scope.go:117] "RemoveContainer" containerID="c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" Mar 17 10:22:44 crc kubenswrapper[4813]: E0317 10:22:44.517895 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:22:52 crc kubenswrapper[4813]: I0317 10:22:52.116209 4813 scope.go:117] "RemoveContainer" containerID="e97424096bd14a70cdf2e53cd4c147471fe3dfd25e61c7104ecb286b2c55d8cd" Mar 17 10:22:56 crc kubenswrapper[4813]: I0317 10:22:56.734959 4813 scope.go:117] "RemoveContainer" containerID="c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" Mar 17 10:22:56 crc kubenswrapper[4813]: E0317 10:22:56.736053 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:23:04 crc kubenswrapper[4813]: I0317 10:23:04.738251 4813 generic.go:334] "Generic (PLEG): container finished" podID="397827d2-1331-46ef-8d9f-15f2ff101941" containerID="2d0f886accd63f5c18fd503b99df71bf14c6f1dffbcb8b80746726a1a296a767" exitCode=0 Mar 17 10:23:04 crc kubenswrapper[4813]: I0317 10:23:04.749544 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pzdlg/must-gather-sv27r" event={"ID":"397827d2-1331-46ef-8d9f-15f2ff101941","Type":"ContainerDied","Data":"2d0f886accd63f5c18fd503b99df71bf14c6f1dffbcb8b80746726a1a296a767"} Mar 17 10:23:04 crc kubenswrapper[4813]: I0317 10:23:04.752078 4813 scope.go:117] "RemoveContainer" containerID="2d0f886accd63f5c18fd503b99df71bf14c6f1dffbcb8b80746726a1a296a767" Mar 17 10:23:05 crc kubenswrapper[4813]: I0317 10:23:05.207463 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pzdlg_must-gather-sv27r_397827d2-1331-46ef-8d9f-15f2ff101941/gather/0.log" Mar 17 10:23:09 crc kubenswrapper[4813]: I0317 10:23:09.731445 4813 scope.go:117] "RemoveContainer" containerID="c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" Mar 17 10:23:09 crc kubenswrapper[4813]: E0317 10:23:09.732767 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:23:16 crc kubenswrapper[4813]: I0317 10:23:16.198079 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pzdlg/must-gather-sv27r"] Mar 17 10:23:16 crc kubenswrapper[4813]: I0317 10:23:16.199095 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-pzdlg/must-gather-sv27r" podUID="397827d2-1331-46ef-8d9f-15f2ff101941" containerName="copy" containerID="cri-o://b6aaa4a49e47badfc0c2c7f28559dcdefbafad39fc8116a1fc85e0b45154ed02" gracePeriod=2 Mar 17 10:23:16 crc kubenswrapper[4813]: I0317 10:23:16.212377 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pzdlg/must-gather-sv27r"] Mar 17 10:23:16 crc kubenswrapper[4813]: I0317 10:23:16.755209 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pzdlg_must-gather-sv27r_397827d2-1331-46ef-8d9f-15f2ff101941/copy/0.log" Mar 17 10:23:16 crc kubenswrapper[4813]: I0317 10:23:16.755964 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pzdlg/must-gather-sv27r" Mar 17 10:23:16 crc kubenswrapper[4813]: I0317 10:23:16.870954 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/397827d2-1331-46ef-8d9f-15f2ff101941-must-gather-output\") pod \"397827d2-1331-46ef-8d9f-15f2ff101941\" (UID: \"397827d2-1331-46ef-8d9f-15f2ff101941\") " Mar 17 10:23:16 crc kubenswrapper[4813]: I0317 10:23:16.871359 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tzzmt\" (UniqueName: \"kubernetes.io/projected/397827d2-1331-46ef-8d9f-15f2ff101941-kube-api-access-tzzmt\") pod \"397827d2-1331-46ef-8d9f-15f2ff101941\" (UID: \"397827d2-1331-46ef-8d9f-15f2ff101941\") " Mar 17 10:23:16 crc kubenswrapper[4813]: I0317 10:23:16.876877 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/397827d2-1331-46ef-8d9f-15f2ff101941-kube-api-access-tzzmt" (OuterVolumeSpecName: "kube-api-access-tzzmt") pod "397827d2-1331-46ef-8d9f-15f2ff101941" (UID: "397827d2-1331-46ef-8d9f-15f2ff101941"). InnerVolumeSpecName "kube-api-access-tzzmt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:23:16 crc kubenswrapper[4813]: I0317 10:23:16.878014 4813 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pzdlg_must-gather-sv27r_397827d2-1331-46ef-8d9f-15f2ff101941/copy/0.log" Mar 17 10:23:16 crc kubenswrapper[4813]: I0317 10:23:16.878299 4813 generic.go:334] "Generic (PLEG): container finished" podID="397827d2-1331-46ef-8d9f-15f2ff101941" containerID="b6aaa4a49e47badfc0c2c7f28559dcdefbafad39fc8116a1fc85e0b45154ed02" exitCode=143 Mar 17 10:23:16 crc kubenswrapper[4813]: I0317 10:23:16.878350 4813 scope.go:117] "RemoveContainer" containerID="b6aaa4a49e47badfc0c2c7f28559dcdefbafad39fc8116a1fc85e0b45154ed02" Mar 17 10:23:16 crc kubenswrapper[4813]: I0317 10:23:16.878351 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pzdlg/must-gather-sv27r" Mar 17 10:23:16 crc kubenswrapper[4813]: I0317 10:23:16.960793 4813 scope.go:117] "RemoveContainer" containerID="2d0f886accd63f5c18fd503b99df71bf14c6f1dffbcb8b80746726a1a296a767" Mar 17 10:23:16 crc kubenswrapper[4813]: I0317 10:23:16.976196 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tzzmt\" (UniqueName: \"kubernetes.io/projected/397827d2-1331-46ef-8d9f-15f2ff101941-kube-api-access-tzzmt\") on node \"crc\" DevicePath \"\"" Mar 17 10:23:17 crc kubenswrapper[4813]: I0317 10:23:17.050227 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/397827d2-1331-46ef-8d9f-15f2ff101941-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "397827d2-1331-46ef-8d9f-15f2ff101941" (UID: "397827d2-1331-46ef-8d9f-15f2ff101941"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:23:17 crc kubenswrapper[4813]: I0317 10:23:17.060585 4813 scope.go:117] "RemoveContainer" containerID="b6aaa4a49e47badfc0c2c7f28559dcdefbafad39fc8116a1fc85e0b45154ed02" Mar 17 10:23:17 crc kubenswrapper[4813]: E0317 10:23:17.061869 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b6aaa4a49e47badfc0c2c7f28559dcdefbafad39fc8116a1fc85e0b45154ed02\": container with ID starting with b6aaa4a49e47badfc0c2c7f28559dcdefbafad39fc8116a1fc85e0b45154ed02 not found: ID does not exist" containerID="b6aaa4a49e47badfc0c2c7f28559dcdefbafad39fc8116a1fc85e0b45154ed02" Mar 17 10:23:17 crc kubenswrapper[4813]: I0317 10:23:17.061911 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b6aaa4a49e47badfc0c2c7f28559dcdefbafad39fc8116a1fc85e0b45154ed02"} err="failed to get container status \"b6aaa4a49e47badfc0c2c7f28559dcdefbafad39fc8116a1fc85e0b45154ed02\": rpc error: code = NotFound desc = could not find container \"b6aaa4a49e47badfc0c2c7f28559dcdefbafad39fc8116a1fc85e0b45154ed02\": container with ID starting with b6aaa4a49e47badfc0c2c7f28559dcdefbafad39fc8116a1fc85e0b45154ed02 not found: ID does not exist" Mar 17 10:23:17 crc kubenswrapper[4813]: I0317 10:23:17.061941 4813 scope.go:117] "RemoveContainer" containerID="2d0f886accd63f5c18fd503b99df71bf14c6f1dffbcb8b80746726a1a296a767" Mar 17 10:23:17 crc kubenswrapper[4813]: E0317 10:23:17.062445 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d0f886accd63f5c18fd503b99df71bf14c6f1dffbcb8b80746726a1a296a767\": container with ID starting with 2d0f886accd63f5c18fd503b99df71bf14c6f1dffbcb8b80746726a1a296a767 not found: ID does not exist" containerID="2d0f886accd63f5c18fd503b99df71bf14c6f1dffbcb8b80746726a1a296a767" Mar 17 10:23:17 crc kubenswrapper[4813]: I0317 10:23:17.062471 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d0f886accd63f5c18fd503b99df71bf14c6f1dffbcb8b80746726a1a296a767"} err="failed to get container status \"2d0f886accd63f5c18fd503b99df71bf14c6f1dffbcb8b80746726a1a296a767\": rpc error: code = NotFound desc = could not find container \"2d0f886accd63f5c18fd503b99df71bf14c6f1dffbcb8b80746726a1a296a767\": container with ID starting with 2d0f886accd63f5c18fd503b99df71bf14c6f1dffbcb8b80746726a1a296a767 not found: ID does not exist" Mar 17 10:23:17 crc kubenswrapper[4813]: I0317 10:23:17.077975 4813 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/397827d2-1331-46ef-8d9f-15f2ff101941-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 17 10:23:18 crc kubenswrapper[4813]: I0317 10:23:18.753954 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="397827d2-1331-46ef-8d9f-15f2ff101941" path="/var/lib/kubelet/pods/397827d2-1331-46ef-8d9f-15f2ff101941/volumes" Mar 17 10:23:24 crc kubenswrapper[4813]: I0317 10:23:24.741710 4813 scope.go:117] "RemoveContainer" containerID="c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" Mar 17 10:23:24 crc kubenswrapper[4813]: E0317 10:23:24.742561 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:23:39 crc kubenswrapper[4813]: I0317 10:23:39.731378 4813 scope.go:117] "RemoveContainer" containerID="c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" Mar 17 10:23:39 crc kubenswrapper[4813]: E0317 10:23:39.732756 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:23:52 crc kubenswrapper[4813]: I0317 10:23:52.894324 4813 scope.go:117] "RemoveContainer" containerID="1c5174094124ee6445542c7c1e2c617e783ffebef96e315ce1532dc37de6b0ce" Mar 17 10:23:54 crc kubenswrapper[4813]: I0317 10:23:54.731516 4813 scope.go:117] "RemoveContainer" containerID="c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" Mar 17 10:23:54 crc kubenswrapper[4813]: E0317 10:23:54.732584 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:24:00 crc kubenswrapper[4813]: I0317 10:24:00.160854 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562384-6lq9f"] Mar 17 10:24:00 crc kubenswrapper[4813]: E0317 10:24:00.161993 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="397827d2-1331-46ef-8d9f-15f2ff101941" containerName="gather" Mar 17 10:24:00 crc kubenswrapper[4813]: I0317 10:24:00.162015 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="397827d2-1331-46ef-8d9f-15f2ff101941" containerName="gather" Mar 17 10:24:00 crc kubenswrapper[4813]: E0317 10:24:00.162036 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca931d57-2657-4c46-bb87-4e3fa445f2ed" containerName="oc" Mar 17 10:24:00 crc kubenswrapper[4813]: I0317 10:24:00.162046 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca931d57-2657-4c46-bb87-4e3fa445f2ed" containerName="oc" Mar 17 10:24:00 crc kubenswrapper[4813]: E0317 10:24:00.162067 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="397827d2-1331-46ef-8d9f-15f2ff101941" containerName="copy" Mar 17 10:24:00 crc kubenswrapper[4813]: I0317 10:24:00.162076 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="397827d2-1331-46ef-8d9f-15f2ff101941" containerName="copy" Mar 17 10:24:00 crc kubenswrapper[4813]: I0317 10:24:00.162354 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca931d57-2657-4c46-bb87-4e3fa445f2ed" containerName="oc" Mar 17 10:24:00 crc kubenswrapper[4813]: I0317 10:24:00.162382 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="397827d2-1331-46ef-8d9f-15f2ff101941" containerName="copy" Mar 17 10:24:00 crc kubenswrapper[4813]: I0317 10:24:00.162410 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="397827d2-1331-46ef-8d9f-15f2ff101941" containerName="gather" Mar 17 10:24:00 crc kubenswrapper[4813]: I0317 10:24:00.163250 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562384-6lq9f" Mar 17 10:24:00 crc kubenswrapper[4813]: I0317 10:24:00.166049 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 10:24:00 crc kubenswrapper[4813]: I0317 10:24:00.167684 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 10:24:00 crc kubenswrapper[4813]: I0317 10:24:00.167945 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 10:24:00 crc kubenswrapper[4813]: I0317 10:24:00.172186 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562384-6lq9f"] Mar 17 10:24:00 crc kubenswrapper[4813]: I0317 10:24:00.328184 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9bfl\" (UniqueName: \"kubernetes.io/projected/35d7a8e4-0ee5-4e66-95f4-36e7a3215a50-kube-api-access-f9bfl\") pod \"auto-csr-approver-29562384-6lq9f\" (UID: \"35d7a8e4-0ee5-4e66-95f4-36e7a3215a50\") " pod="openshift-infra/auto-csr-approver-29562384-6lq9f" Mar 17 10:24:00 crc kubenswrapper[4813]: I0317 10:24:00.430456 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9bfl\" (UniqueName: \"kubernetes.io/projected/35d7a8e4-0ee5-4e66-95f4-36e7a3215a50-kube-api-access-f9bfl\") pod \"auto-csr-approver-29562384-6lq9f\" (UID: \"35d7a8e4-0ee5-4e66-95f4-36e7a3215a50\") " pod="openshift-infra/auto-csr-approver-29562384-6lq9f" Mar 17 10:24:00 crc kubenswrapper[4813]: I0317 10:24:00.451938 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9bfl\" (UniqueName: \"kubernetes.io/projected/35d7a8e4-0ee5-4e66-95f4-36e7a3215a50-kube-api-access-f9bfl\") pod \"auto-csr-approver-29562384-6lq9f\" (UID: \"35d7a8e4-0ee5-4e66-95f4-36e7a3215a50\") " pod="openshift-infra/auto-csr-approver-29562384-6lq9f" Mar 17 10:24:00 crc kubenswrapper[4813]: I0317 10:24:00.490077 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562384-6lq9f" Mar 17 10:24:00 crc kubenswrapper[4813]: I0317 10:24:00.987990 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562384-6lq9f"] Mar 17 10:24:01 crc kubenswrapper[4813]: I0317 10:24:01.347789 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562384-6lq9f" event={"ID":"35d7a8e4-0ee5-4e66-95f4-36e7a3215a50","Type":"ContainerStarted","Data":"65c0a22c82e3566fa0f6e181b5dc10f049d6055f4cca2e09d1e59a2b0a675e7f"} Mar 17 10:24:02 crc kubenswrapper[4813]: I0317 10:24:02.361431 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562384-6lq9f" event={"ID":"35d7a8e4-0ee5-4e66-95f4-36e7a3215a50","Type":"ContainerStarted","Data":"825520d26b647ca070d8d1c11b05bf14a6458b9379c6bd7b3edced831251617c"} Mar 17 10:24:02 crc kubenswrapper[4813]: I0317 10:24:02.380947 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29562384-6lq9f" podStartSLOduration=1.5529700339999999 podStartE2EDuration="2.38091645s" podCreationTimestamp="2026-03-17 10:24:00 +0000 UTC" firstStartedPulling="2026-03-17 10:24:00.990147629 +0000 UTC m=+4463.090951128" lastFinishedPulling="2026-03-17 10:24:01.818094005 +0000 UTC m=+4463.918897544" observedRunningTime="2026-03-17 10:24:02.376301926 +0000 UTC m=+4464.477105475" watchObservedRunningTime="2026-03-17 10:24:02.38091645 +0000 UTC m=+4464.481719989" Mar 17 10:24:03 crc kubenswrapper[4813]: I0317 10:24:03.373811 4813 generic.go:334] "Generic (PLEG): container finished" podID="35d7a8e4-0ee5-4e66-95f4-36e7a3215a50" containerID="825520d26b647ca070d8d1c11b05bf14a6458b9379c6bd7b3edced831251617c" exitCode=0 Mar 17 10:24:03 crc kubenswrapper[4813]: I0317 10:24:03.373876 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562384-6lq9f" event={"ID":"35d7a8e4-0ee5-4e66-95f4-36e7a3215a50","Type":"ContainerDied","Data":"825520d26b647ca070d8d1c11b05bf14a6458b9379c6bd7b3edced831251617c"} Mar 17 10:24:04 crc kubenswrapper[4813]: I0317 10:24:04.804697 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562384-6lq9f" Mar 17 10:24:04 crc kubenswrapper[4813]: I0317 10:24:04.836590 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9bfl\" (UniqueName: \"kubernetes.io/projected/35d7a8e4-0ee5-4e66-95f4-36e7a3215a50-kube-api-access-f9bfl\") pod \"35d7a8e4-0ee5-4e66-95f4-36e7a3215a50\" (UID: \"35d7a8e4-0ee5-4e66-95f4-36e7a3215a50\") " Mar 17 10:24:04 crc kubenswrapper[4813]: I0317 10:24:04.853900 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35d7a8e4-0ee5-4e66-95f4-36e7a3215a50-kube-api-access-f9bfl" (OuterVolumeSpecName: "kube-api-access-f9bfl") pod "35d7a8e4-0ee5-4e66-95f4-36e7a3215a50" (UID: "35d7a8e4-0ee5-4e66-95f4-36e7a3215a50"). InnerVolumeSpecName "kube-api-access-f9bfl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:24:04 crc kubenswrapper[4813]: I0317 10:24:04.939519 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9bfl\" (UniqueName: \"kubernetes.io/projected/35d7a8e4-0ee5-4e66-95f4-36e7a3215a50-kube-api-access-f9bfl\") on node \"crc\" DevicePath \"\"" Mar 17 10:24:05 crc kubenswrapper[4813]: I0317 10:24:05.423090 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562384-6lq9f" event={"ID":"35d7a8e4-0ee5-4e66-95f4-36e7a3215a50","Type":"ContainerDied","Data":"65c0a22c82e3566fa0f6e181b5dc10f049d6055f4cca2e09d1e59a2b0a675e7f"} Mar 17 10:24:05 crc kubenswrapper[4813]: I0317 10:24:05.423490 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="65c0a22c82e3566fa0f6e181b5dc10f049d6055f4cca2e09d1e59a2b0a675e7f" Mar 17 10:24:05 crc kubenswrapper[4813]: I0317 10:24:05.423203 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562384-6lq9f" Mar 17 10:24:05 crc kubenswrapper[4813]: I0317 10:24:05.485182 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562378-4ghd7"] Mar 17 10:24:05 crc kubenswrapper[4813]: I0317 10:24:05.495570 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562378-4ghd7"] Mar 17 10:24:06 crc kubenswrapper[4813]: I0317 10:24:06.748552 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e34845ff-00c2-44f9-a3ed-fb16e52e1a06" path="/var/lib/kubelet/pods/e34845ff-00c2-44f9-a3ed-fb16e52e1a06/volumes" Mar 17 10:24:08 crc kubenswrapper[4813]: I0317 10:24:08.746484 4813 scope.go:117] "RemoveContainer" containerID="c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" Mar 17 10:24:08 crc kubenswrapper[4813]: E0317 10:24:08.747075 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:24:23 crc kubenswrapper[4813]: I0317 10:24:23.732360 4813 scope.go:117] "RemoveContainer" containerID="c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" Mar 17 10:24:23 crc kubenswrapper[4813]: E0317 10:24:23.733485 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:24:34 crc kubenswrapper[4813]: I0317 10:24:34.732038 4813 scope.go:117] "RemoveContainer" containerID="c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" Mar 17 10:24:34 crc kubenswrapper[4813]: E0317 10:24:34.732868 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:24:46 crc kubenswrapper[4813]: I0317 10:24:46.731136 4813 scope.go:117] "RemoveContainer" containerID="c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" Mar 17 10:24:46 crc kubenswrapper[4813]: E0317 10:24:46.732508 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:24:53 crc kubenswrapper[4813]: I0317 10:24:53.010829 4813 scope.go:117] "RemoveContainer" containerID="cc3051458887669a97f7638605718a52e19ac68b4bff84fa90ef107aa3d50d8e" Mar 17 10:25:01 crc kubenswrapper[4813]: I0317 10:25:01.731304 4813 scope.go:117] "RemoveContainer" containerID="c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" Mar 17 10:25:01 crc kubenswrapper[4813]: E0317 10:25:01.732276 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:25:15 crc kubenswrapper[4813]: I0317 10:25:15.731542 4813 scope.go:117] "RemoveContainer" containerID="c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" Mar 17 10:25:15 crc kubenswrapper[4813]: E0317 10:25:15.732688 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:25:26 crc kubenswrapper[4813]: I0317 10:25:26.731695 4813 scope.go:117] "RemoveContainer" containerID="c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" Mar 17 10:25:26 crc kubenswrapper[4813]: E0317 10:25:26.732698 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:25:39 crc kubenswrapper[4813]: I0317 10:25:39.730510 4813 scope.go:117] "RemoveContainer" containerID="c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" Mar 17 10:25:39 crc kubenswrapper[4813]: E0317 10:25:39.731429 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:25:52 crc kubenswrapper[4813]: I0317 10:25:52.731776 4813 scope.go:117] "RemoveContainer" containerID="c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" Mar 17 10:25:52 crc kubenswrapper[4813]: E0317 10:25:52.732939 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:26:00 crc kubenswrapper[4813]: I0317 10:26:00.171901 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29562386-qfdsl"] Mar 17 10:26:00 crc kubenswrapper[4813]: E0317 10:26:00.173003 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35d7a8e4-0ee5-4e66-95f4-36e7a3215a50" containerName="oc" Mar 17 10:26:00 crc kubenswrapper[4813]: I0317 10:26:00.173019 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="35d7a8e4-0ee5-4e66-95f4-36e7a3215a50" containerName="oc" Mar 17 10:26:00 crc kubenswrapper[4813]: I0317 10:26:00.173460 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="35d7a8e4-0ee5-4e66-95f4-36e7a3215a50" containerName="oc" Mar 17 10:26:00 crc kubenswrapper[4813]: I0317 10:26:00.174442 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562386-qfdsl" Mar 17 10:26:00 crc kubenswrapper[4813]: I0317 10:26:00.180394 4813 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-wsvtm" Mar 17 10:26:00 crc kubenswrapper[4813]: I0317 10:26:00.181168 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 17 10:26:00 crc kubenswrapper[4813]: I0317 10:26:00.181274 4813 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 17 10:26:00 crc kubenswrapper[4813]: I0317 10:26:00.219246 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562386-qfdsl"] Mar 17 10:26:00 crc kubenswrapper[4813]: I0317 10:26:00.294737 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zdd6h\" (UniqueName: \"kubernetes.io/projected/5f4f2f0d-be75-43c5-a1c3-a55d71d17684-kube-api-access-zdd6h\") pod \"auto-csr-approver-29562386-qfdsl\" (UID: \"5f4f2f0d-be75-43c5-a1c3-a55d71d17684\") " pod="openshift-infra/auto-csr-approver-29562386-qfdsl" Mar 17 10:26:00 crc kubenswrapper[4813]: I0317 10:26:00.397377 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zdd6h\" (UniqueName: \"kubernetes.io/projected/5f4f2f0d-be75-43c5-a1c3-a55d71d17684-kube-api-access-zdd6h\") pod \"auto-csr-approver-29562386-qfdsl\" (UID: \"5f4f2f0d-be75-43c5-a1c3-a55d71d17684\") " pod="openshift-infra/auto-csr-approver-29562386-qfdsl" Mar 17 10:26:00 crc kubenswrapper[4813]: I0317 10:26:00.424763 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zdd6h\" (UniqueName: \"kubernetes.io/projected/5f4f2f0d-be75-43c5-a1c3-a55d71d17684-kube-api-access-zdd6h\") pod \"auto-csr-approver-29562386-qfdsl\" (UID: \"5f4f2f0d-be75-43c5-a1c3-a55d71d17684\") " pod="openshift-infra/auto-csr-approver-29562386-qfdsl" Mar 17 10:26:00 crc kubenswrapper[4813]: I0317 10:26:00.521863 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562386-qfdsl" Mar 17 10:26:01 crc kubenswrapper[4813]: I0317 10:26:01.003238 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29562386-qfdsl"] Mar 17 10:26:01 crc kubenswrapper[4813]: I0317 10:26:01.779211 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562386-qfdsl" event={"ID":"5f4f2f0d-be75-43c5-a1c3-a55d71d17684","Type":"ContainerStarted","Data":"218eba75831274f0700c1b9507803730e1f8f57442e6540a6bc9a11e5a5de8a8"} Mar 17 10:26:03 crc kubenswrapper[4813]: I0317 10:26:03.732092 4813 scope.go:117] "RemoveContainer" containerID="c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" Mar 17 10:26:03 crc kubenswrapper[4813]: E0317 10:26:03.733403 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:26:03 crc kubenswrapper[4813]: I0317 10:26:03.822439 4813 generic.go:334] "Generic (PLEG): container finished" podID="5f4f2f0d-be75-43c5-a1c3-a55d71d17684" containerID="f4c4bfe99f93a12c37d076be4a50fab7c9a46b7e0f1dbaff877f5762ad2f805f" exitCode=0 Mar 17 10:26:03 crc kubenswrapper[4813]: I0317 10:26:03.822504 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562386-qfdsl" event={"ID":"5f4f2f0d-be75-43c5-a1c3-a55d71d17684","Type":"ContainerDied","Data":"f4c4bfe99f93a12c37d076be4a50fab7c9a46b7e0f1dbaff877f5762ad2f805f"} Mar 17 10:26:05 crc kubenswrapper[4813]: I0317 10:26:05.293749 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562386-qfdsl" Mar 17 10:26:05 crc kubenswrapper[4813]: I0317 10:26:05.313717 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zdd6h\" (UniqueName: \"kubernetes.io/projected/5f4f2f0d-be75-43c5-a1c3-a55d71d17684-kube-api-access-zdd6h\") pod \"5f4f2f0d-be75-43c5-a1c3-a55d71d17684\" (UID: \"5f4f2f0d-be75-43c5-a1c3-a55d71d17684\") " Mar 17 10:26:05 crc kubenswrapper[4813]: I0317 10:26:05.320523 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f4f2f0d-be75-43c5-a1c3-a55d71d17684-kube-api-access-zdd6h" (OuterVolumeSpecName: "kube-api-access-zdd6h") pod "5f4f2f0d-be75-43c5-a1c3-a55d71d17684" (UID: "5f4f2f0d-be75-43c5-a1c3-a55d71d17684"). InnerVolumeSpecName "kube-api-access-zdd6h". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:26:05 crc kubenswrapper[4813]: I0317 10:26:05.416730 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zdd6h\" (UniqueName: \"kubernetes.io/projected/5f4f2f0d-be75-43c5-a1c3-a55d71d17684-kube-api-access-zdd6h\") on node \"crc\" DevicePath \"\"" Mar 17 10:26:05 crc kubenswrapper[4813]: I0317 10:26:05.849454 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29562386-qfdsl" event={"ID":"5f4f2f0d-be75-43c5-a1c3-a55d71d17684","Type":"ContainerDied","Data":"218eba75831274f0700c1b9507803730e1f8f57442e6540a6bc9a11e5a5de8a8"} Mar 17 10:26:05 crc kubenswrapper[4813]: I0317 10:26:05.849905 4813 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="218eba75831274f0700c1b9507803730e1f8f57442e6540a6bc9a11e5a5de8a8" Mar 17 10:26:05 crc kubenswrapper[4813]: I0317 10:26:05.849532 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29562386-qfdsl" Mar 17 10:26:06 crc kubenswrapper[4813]: I0317 10:26:06.447939 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29562380-p499h"] Mar 17 10:26:06 crc kubenswrapper[4813]: I0317 10:26:06.457627 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29562380-p499h"] Mar 17 10:26:06 crc kubenswrapper[4813]: I0317 10:26:06.743656 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf51d593-228d-4052-aa60-54849843ee9a" path="/var/lib/kubelet/pods/bf51d593-228d-4052-aa60-54849843ee9a/volumes" Mar 17 10:26:17 crc kubenswrapper[4813]: I0317 10:26:17.730354 4813 scope.go:117] "RemoveContainer" containerID="c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" Mar 17 10:26:17 crc kubenswrapper[4813]: E0317 10:26:17.731162 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:26:22 crc kubenswrapper[4813]: I0317 10:26:22.269495 4813 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-pzfmn"] Mar 17 10:26:22 crc kubenswrapper[4813]: E0317 10:26:22.270683 4813 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f4f2f0d-be75-43c5-a1c3-a55d71d17684" containerName="oc" Mar 17 10:26:22 crc kubenswrapper[4813]: I0317 10:26:22.270724 4813 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f4f2f0d-be75-43c5-a1c3-a55d71d17684" containerName="oc" Mar 17 10:26:22 crc kubenswrapper[4813]: I0317 10:26:22.271087 4813 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f4f2f0d-be75-43c5-a1c3-a55d71d17684" containerName="oc" Mar 17 10:26:22 crc kubenswrapper[4813]: I0317 10:26:22.273024 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pzfmn" Mar 17 10:26:22 crc kubenswrapper[4813]: I0317 10:26:22.305860 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pzfmn"] Mar 17 10:26:22 crc kubenswrapper[4813]: I0317 10:26:22.384491 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqpzb\" (UniqueName: \"kubernetes.io/projected/ca0ea2ee-6a2d-4884-9696-ecadd248f18f-kube-api-access-pqpzb\") pod \"redhat-operators-pzfmn\" (UID: \"ca0ea2ee-6a2d-4884-9696-ecadd248f18f\") " pod="openshift-marketplace/redhat-operators-pzfmn" Mar 17 10:26:22 crc kubenswrapper[4813]: I0317 10:26:22.384538 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca0ea2ee-6a2d-4884-9696-ecadd248f18f-catalog-content\") pod \"redhat-operators-pzfmn\" (UID: \"ca0ea2ee-6a2d-4884-9696-ecadd248f18f\") " pod="openshift-marketplace/redhat-operators-pzfmn" Mar 17 10:26:22 crc kubenswrapper[4813]: I0317 10:26:22.386782 4813 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca0ea2ee-6a2d-4884-9696-ecadd248f18f-utilities\") pod \"redhat-operators-pzfmn\" (UID: \"ca0ea2ee-6a2d-4884-9696-ecadd248f18f\") " pod="openshift-marketplace/redhat-operators-pzfmn" Mar 17 10:26:22 crc kubenswrapper[4813]: I0317 10:26:22.488982 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqpzb\" (UniqueName: \"kubernetes.io/projected/ca0ea2ee-6a2d-4884-9696-ecadd248f18f-kube-api-access-pqpzb\") pod \"redhat-operators-pzfmn\" (UID: \"ca0ea2ee-6a2d-4884-9696-ecadd248f18f\") " pod="openshift-marketplace/redhat-operators-pzfmn" Mar 17 10:26:22 crc kubenswrapper[4813]: I0317 10:26:22.489033 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca0ea2ee-6a2d-4884-9696-ecadd248f18f-catalog-content\") pod \"redhat-operators-pzfmn\" (UID: \"ca0ea2ee-6a2d-4884-9696-ecadd248f18f\") " pod="openshift-marketplace/redhat-operators-pzfmn" Mar 17 10:26:22 crc kubenswrapper[4813]: I0317 10:26:22.489081 4813 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca0ea2ee-6a2d-4884-9696-ecadd248f18f-utilities\") pod \"redhat-operators-pzfmn\" (UID: \"ca0ea2ee-6a2d-4884-9696-ecadd248f18f\") " pod="openshift-marketplace/redhat-operators-pzfmn" Mar 17 10:26:22 crc kubenswrapper[4813]: I0317 10:26:22.489678 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca0ea2ee-6a2d-4884-9696-ecadd248f18f-catalog-content\") pod \"redhat-operators-pzfmn\" (UID: \"ca0ea2ee-6a2d-4884-9696-ecadd248f18f\") " pod="openshift-marketplace/redhat-operators-pzfmn" Mar 17 10:26:22 crc kubenswrapper[4813]: I0317 10:26:22.490379 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca0ea2ee-6a2d-4884-9696-ecadd248f18f-utilities\") pod \"redhat-operators-pzfmn\" (UID: \"ca0ea2ee-6a2d-4884-9696-ecadd248f18f\") " pod="openshift-marketplace/redhat-operators-pzfmn" Mar 17 10:26:22 crc kubenswrapper[4813]: I0317 10:26:22.510364 4813 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqpzb\" (UniqueName: \"kubernetes.io/projected/ca0ea2ee-6a2d-4884-9696-ecadd248f18f-kube-api-access-pqpzb\") pod \"redhat-operators-pzfmn\" (UID: \"ca0ea2ee-6a2d-4884-9696-ecadd248f18f\") " pod="openshift-marketplace/redhat-operators-pzfmn" Mar 17 10:26:22 crc kubenswrapper[4813]: I0317 10:26:22.606987 4813 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pzfmn" Mar 17 10:26:23 crc kubenswrapper[4813]: W0317 10:26:23.109861 4813 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podca0ea2ee_6a2d_4884_9696_ecadd248f18f.slice/crio-67d3f731bd6c90749aafc48b12396c23f33775e6aac796b47bbca527bc114569 WatchSource:0}: Error finding container 67d3f731bd6c90749aafc48b12396c23f33775e6aac796b47bbca527bc114569: Status 404 returned error can't find the container with id 67d3f731bd6c90749aafc48b12396c23f33775e6aac796b47bbca527bc114569 Mar 17 10:26:23 crc kubenswrapper[4813]: I0317 10:26:23.111701 4813 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-pzfmn"] Mar 17 10:26:24 crc kubenswrapper[4813]: I0317 10:26:24.058088 4813 generic.go:334] "Generic (PLEG): container finished" podID="ca0ea2ee-6a2d-4884-9696-ecadd248f18f" containerID="7b98f73d433f0e8b3403b745dc4575cab8090d6df939d035272008a1e8411450" exitCode=0 Mar 17 10:26:24 crc kubenswrapper[4813]: I0317 10:26:24.058254 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pzfmn" event={"ID":"ca0ea2ee-6a2d-4884-9696-ecadd248f18f","Type":"ContainerDied","Data":"7b98f73d433f0e8b3403b745dc4575cab8090d6df939d035272008a1e8411450"} Mar 17 10:26:24 crc kubenswrapper[4813]: I0317 10:26:24.058554 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pzfmn" event={"ID":"ca0ea2ee-6a2d-4884-9696-ecadd248f18f","Type":"ContainerStarted","Data":"67d3f731bd6c90749aafc48b12396c23f33775e6aac796b47bbca527bc114569"} Mar 17 10:26:26 crc kubenswrapper[4813]: I0317 10:26:26.079930 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pzfmn" event={"ID":"ca0ea2ee-6a2d-4884-9696-ecadd248f18f","Type":"ContainerStarted","Data":"a60adbcb5eb1f3a38423f5fbc365280949e3d3b3e5bb764298daec395cef8911"} Mar 17 10:26:29 crc kubenswrapper[4813]: I0317 10:26:29.107653 4813 generic.go:334] "Generic (PLEG): container finished" podID="ca0ea2ee-6a2d-4884-9696-ecadd248f18f" containerID="a60adbcb5eb1f3a38423f5fbc365280949e3d3b3e5bb764298daec395cef8911" exitCode=0 Mar 17 10:26:29 crc kubenswrapper[4813]: I0317 10:26:29.107768 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pzfmn" event={"ID":"ca0ea2ee-6a2d-4884-9696-ecadd248f18f","Type":"ContainerDied","Data":"a60adbcb5eb1f3a38423f5fbc365280949e3d3b3e5bb764298daec395cef8911"} Mar 17 10:26:31 crc kubenswrapper[4813]: I0317 10:26:31.128981 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pzfmn" event={"ID":"ca0ea2ee-6a2d-4884-9696-ecadd248f18f","Type":"ContainerStarted","Data":"ec2d1607babe21c7f5b5ff3dbd21a466931394aa3e60000eb65366764d8b7a3f"} Mar 17 10:26:31 crc kubenswrapper[4813]: I0317 10:26:31.149139 4813 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-pzfmn" podStartSLOduration=3.652578879 podStartE2EDuration="9.149095813s" podCreationTimestamp="2026-03-17 10:26:22 +0000 UTC" firstStartedPulling="2026-03-17 10:26:24.059736747 +0000 UTC m=+4606.160540256" lastFinishedPulling="2026-03-17 10:26:29.556253691 +0000 UTC m=+4611.657057190" observedRunningTime="2026-03-17 10:26:31.148650069 +0000 UTC m=+4613.249453588" watchObservedRunningTime="2026-03-17 10:26:31.149095813 +0000 UTC m=+4613.249899312" Mar 17 10:26:32 crc kubenswrapper[4813]: I0317 10:26:32.609680 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-pzfmn" Mar 17 10:26:32 crc kubenswrapper[4813]: I0317 10:26:32.610151 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-pzfmn" Mar 17 10:26:32 crc kubenswrapper[4813]: I0317 10:26:32.731185 4813 scope.go:117] "RemoveContainer" containerID="c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" Mar 17 10:26:32 crc kubenswrapper[4813]: E0317 10:26:32.731845 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:26:33 crc kubenswrapper[4813]: I0317 10:26:33.687105 4813 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-pzfmn" podUID="ca0ea2ee-6a2d-4884-9696-ecadd248f18f" containerName="registry-server" probeResult="failure" output=< Mar 17 10:26:33 crc kubenswrapper[4813]: timeout: failed to connect service ":50051" within 1s Mar 17 10:26:33 crc kubenswrapper[4813]: > Mar 17 10:26:42 crc kubenswrapper[4813]: I0317 10:26:42.928849 4813 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-pzfmn" Mar 17 10:26:42 crc kubenswrapper[4813]: I0317 10:26:42.989707 4813 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-pzfmn" Mar 17 10:26:43 crc kubenswrapper[4813]: I0317 10:26:43.175476 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pzfmn"] Mar 17 10:26:44 crc kubenswrapper[4813]: I0317 10:26:44.258918 4813 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-pzfmn" podUID="ca0ea2ee-6a2d-4884-9696-ecadd248f18f" containerName="registry-server" containerID="cri-o://ec2d1607babe21c7f5b5ff3dbd21a466931394aa3e60000eb65366764d8b7a3f" gracePeriod=2 Mar 17 10:26:44 crc kubenswrapper[4813]: I0317 10:26:44.730265 4813 scope.go:117] "RemoveContainer" containerID="c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" Mar 17 10:26:44 crc kubenswrapper[4813]: E0317 10:26:44.730836 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" Mar 17 10:26:44 crc kubenswrapper[4813]: I0317 10:26:44.886515 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pzfmn" Mar 17 10:26:44 crc kubenswrapper[4813]: I0317 10:26:44.989945 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca0ea2ee-6a2d-4884-9696-ecadd248f18f-catalog-content\") pod \"ca0ea2ee-6a2d-4884-9696-ecadd248f18f\" (UID: \"ca0ea2ee-6a2d-4884-9696-ecadd248f18f\") " Mar 17 10:26:44 crc kubenswrapper[4813]: I0317 10:26:44.990113 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqpzb\" (UniqueName: \"kubernetes.io/projected/ca0ea2ee-6a2d-4884-9696-ecadd248f18f-kube-api-access-pqpzb\") pod \"ca0ea2ee-6a2d-4884-9696-ecadd248f18f\" (UID: \"ca0ea2ee-6a2d-4884-9696-ecadd248f18f\") " Mar 17 10:26:44 crc kubenswrapper[4813]: I0317 10:26:44.990149 4813 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca0ea2ee-6a2d-4884-9696-ecadd248f18f-utilities\") pod \"ca0ea2ee-6a2d-4884-9696-ecadd248f18f\" (UID: \"ca0ea2ee-6a2d-4884-9696-ecadd248f18f\") " Mar 17 10:26:44 crc kubenswrapper[4813]: I0317 10:26:44.991115 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca0ea2ee-6a2d-4884-9696-ecadd248f18f-utilities" (OuterVolumeSpecName: "utilities") pod "ca0ea2ee-6a2d-4884-9696-ecadd248f18f" (UID: "ca0ea2ee-6a2d-4884-9696-ecadd248f18f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:26:44 crc kubenswrapper[4813]: I0317 10:26:44.996733 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca0ea2ee-6a2d-4884-9696-ecadd248f18f-kube-api-access-pqpzb" (OuterVolumeSpecName: "kube-api-access-pqpzb") pod "ca0ea2ee-6a2d-4884-9696-ecadd248f18f" (UID: "ca0ea2ee-6a2d-4884-9696-ecadd248f18f"). InnerVolumeSpecName "kube-api-access-pqpzb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 17 10:26:45 crc kubenswrapper[4813]: I0317 10:26:45.093060 4813 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqpzb\" (UniqueName: \"kubernetes.io/projected/ca0ea2ee-6a2d-4884-9696-ecadd248f18f-kube-api-access-pqpzb\") on node \"crc\" DevicePath \"\"" Mar 17 10:26:45 crc kubenswrapper[4813]: I0317 10:26:45.093109 4813 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca0ea2ee-6a2d-4884-9696-ecadd248f18f-utilities\") on node \"crc\" DevicePath \"\"" Mar 17 10:26:45 crc kubenswrapper[4813]: I0317 10:26:45.124434 4813 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca0ea2ee-6a2d-4884-9696-ecadd248f18f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ca0ea2ee-6a2d-4884-9696-ecadd248f18f" (UID: "ca0ea2ee-6a2d-4884-9696-ecadd248f18f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 17 10:26:45 crc kubenswrapper[4813]: I0317 10:26:45.194341 4813 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca0ea2ee-6a2d-4884-9696-ecadd248f18f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 17 10:26:45 crc kubenswrapper[4813]: I0317 10:26:45.274876 4813 generic.go:334] "Generic (PLEG): container finished" podID="ca0ea2ee-6a2d-4884-9696-ecadd248f18f" containerID="ec2d1607babe21c7f5b5ff3dbd21a466931394aa3e60000eb65366764d8b7a3f" exitCode=0 Mar 17 10:26:45 crc kubenswrapper[4813]: I0317 10:26:45.274932 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pzfmn" event={"ID":"ca0ea2ee-6a2d-4884-9696-ecadd248f18f","Type":"ContainerDied","Data":"ec2d1607babe21c7f5b5ff3dbd21a466931394aa3e60000eb65366764d8b7a3f"} Mar 17 10:26:45 crc kubenswrapper[4813]: I0317 10:26:45.274990 4813 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-pzfmn" Mar 17 10:26:45 crc kubenswrapper[4813]: I0317 10:26:45.274968 4813 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-pzfmn" event={"ID":"ca0ea2ee-6a2d-4884-9696-ecadd248f18f","Type":"ContainerDied","Data":"67d3f731bd6c90749aafc48b12396c23f33775e6aac796b47bbca527bc114569"} Mar 17 10:26:45 crc kubenswrapper[4813]: I0317 10:26:45.275039 4813 scope.go:117] "RemoveContainer" containerID="ec2d1607babe21c7f5b5ff3dbd21a466931394aa3e60000eb65366764d8b7a3f" Mar 17 10:26:45 crc kubenswrapper[4813]: I0317 10:26:45.314115 4813 scope.go:117] "RemoveContainer" containerID="a60adbcb5eb1f3a38423f5fbc365280949e3d3b3e5bb764298daec395cef8911" Mar 17 10:26:45 crc kubenswrapper[4813]: I0317 10:26:45.339757 4813 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-pzfmn"] Mar 17 10:26:45 crc kubenswrapper[4813]: I0317 10:26:45.351302 4813 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-pzfmn"] Mar 17 10:26:45 crc kubenswrapper[4813]: I0317 10:26:45.383018 4813 scope.go:117] "RemoveContainer" containerID="7b98f73d433f0e8b3403b745dc4575cab8090d6df939d035272008a1e8411450" Mar 17 10:26:45 crc kubenswrapper[4813]: I0317 10:26:45.400266 4813 scope.go:117] "RemoveContainer" containerID="ec2d1607babe21c7f5b5ff3dbd21a466931394aa3e60000eb65366764d8b7a3f" Mar 17 10:26:45 crc kubenswrapper[4813]: E0317 10:26:45.401017 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec2d1607babe21c7f5b5ff3dbd21a466931394aa3e60000eb65366764d8b7a3f\": container with ID starting with ec2d1607babe21c7f5b5ff3dbd21a466931394aa3e60000eb65366764d8b7a3f not found: ID does not exist" containerID="ec2d1607babe21c7f5b5ff3dbd21a466931394aa3e60000eb65366764d8b7a3f" Mar 17 10:26:45 crc kubenswrapper[4813]: I0317 10:26:45.401046 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec2d1607babe21c7f5b5ff3dbd21a466931394aa3e60000eb65366764d8b7a3f"} err="failed to get container status \"ec2d1607babe21c7f5b5ff3dbd21a466931394aa3e60000eb65366764d8b7a3f\": rpc error: code = NotFound desc = could not find container \"ec2d1607babe21c7f5b5ff3dbd21a466931394aa3e60000eb65366764d8b7a3f\": container with ID starting with ec2d1607babe21c7f5b5ff3dbd21a466931394aa3e60000eb65366764d8b7a3f not found: ID does not exist" Mar 17 10:26:45 crc kubenswrapper[4813]: I0317 10:26:45.401066 4813 scope.go:117] "RemoveContainer" containerID="a60adbcb5eb1f3a38423f5fbc365280949e3d3b3e5bb764298daec395cef8911" Mar 17 10:26:45 crc kubenswrapper[4813]: E0317 10:26:45.401419 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a60adbcb5eb1f3a38423f5fbc365280949e3d3b3e5bb764298daec395cef8911\": container with ID starting with a60adbcb5eb1f3a38423f5fbc365280949e3d3b3e5bb764298daec395cef8911 not found: ID does not exist" containerID="a60adbcb5eb1f3a38423f5fbc365280949e3d3b3e5bb764298daec395cef8911" Mar 17 10:26:45 crc kubenswrapper[4813]: I0317 10:26:45.401469 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a60adbcb5eb1f3a38423f5fbc365280949e3d3b3e5bb764298daec395cef8911"} err="failed to get container status \"a60adbcb5eb1f3a38423f5fbc365280949e3d3b3e5bb764298daec395cef8911\": rpc error: code = NotFound desc = could not find container \"a60adbcb5eb1f3a38423f5fbc365280949e3d3b3e5bb764298daec395cef8911\": container with ID starting with a60adbcb5eb1f3a38423f5fbc365280949e3d3b3e5bb764298daec395cef8911 not found: ID does not exist" Mar 17 10:26:45 crc kubenswrapper[4813]: I0317 10:26:45.401501 4813 scope.go:117] "RemoveContainer" containerID="7b98f73d433f0e8b3403b745dc4575cab8090d6df939d035272008a1e8411450" Mar 17 10:26:45 crc kubenswrapper[4813]: E0317 10:26:45.401823 4813 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b98f73d433f0e8b3403b745dc4575cab8090d6df939d035272008a1e8411450\": container with ID starting with 7b98f73d433f0e8b3403b745dc4575cab8090d6df939d035272008a1e8411450 not found: ID does not exist" containerID="7b98f73d433f0e8b3403b745dc4575cab8090d6df939d035272008a1e8411450" Mar 17 10:26:45 crc kubenswrapper[4813]: I0317 10:26:45.401847 4813 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b98f73d433f0e8b3403b745dc4575cab8090d6df939d035272008a1e8411450"} err="failed to get container status \"7b98f73d433f0e8b3403b745dc4575cab8090d6df939d035272008a1e8411450\": rpc error: code = NotFound desc = could not find container \"7b98f73d433f0e8b3403b745dc4575cab8090d6df939d035272008a1e8411450\": container with ID starting with 7b98f73d433f0e8b3403b745dc4575cab8090d6df939d035272008a1e8411450 not found: ID does not exist" Mar 17 10:26:46 crc kubenswrapper[4813]: I0317 10:26:46.776317 4813 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca0ea2ee-6a2d-4884-9696-ecadd248f18f" path="/var/lib/kubelet/pods/ca0ea2ee-6a2d-4884-9696-ecadd248f18f/volumes" Mar 17 10:26:53 crc kubenswrapper[4813]: I0317 10:26:53.136784 4813 scope.go:117] "RemoveContainer" containerID="0661df999c2af64696b83686113e55692c266b6d6108c7b5de72bb526c3d38cb" Mar 17 10:26:56 crc kubenswrapper[4813]: I0317 10:26:56.732058 4813 scope.go:117] "RemoveContainer" containerID="c939ae554ba99759aaffa3f4816df16f286148b117d432f9c2e401505f7214ee" Mar 17 10:26:56 crc kubenswrapper[4813]: E0317 10:26:56.734206 4813 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-7xxkf_openshift-machine-config-operator(e7e156ce-b279-4a25-a27e-2da7d0ba6437)\"" pod="openshift-machine-config-operator/machine-config-daemon-7xxkf" podUID="e7e156ce-b279-4a25-a27e-2da7d0ba6437" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515156226200024444 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015156226201017362 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015156214607016514 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015156214607015464 5ustar corecore